remove cache configs in model definitions (#4031)

This commit is contained in:
Lianmin Zheng
2025-03-03 05:00:50 -08:00
committed by GitHub
parent a3ab768a2b
commit 1a8f995c46
4 changed files with 0 additions and 72 deletions

View File

@@ -359,7 +359,6 @@ class Grok1ForCausalLM(nn.Module):
self,
config: PretrainedConfig,
quant_config: Optional[QuantizationConfig] = None,
cache_config=None,
) -> None:
super().__init__()
self.config = config

View File

@@ -106,7 +106,6 @@ class LlamaForCausalLMEagle(LlamaForCausalLM):
self,
config: LlamaConfig,
quant_config: Optional[QuantizationConfig] = None,
cache_config=None,
) -> None:
nn.Module.__init__(self)
self.config = config

View File

@@ -107,7 +107,6 @@ class Qwen2ForCausalLMEagle(Qwen2ForCausalLM):
self,
config: Qwen2Config,
quant_config: Optional[QuantizationConfig] = None,
cache_config=None,
) -> None:
nn.Module.__init__(self)
self.config = config