add deepseekv3 and llama4

This commit is contained in:
Chranos
2026-02-11 14:39:48 +08:00
parent d93c740e4d
commit d9d1f3a724

View File

@@ -20,7 +20,7 @@ def set_default_torch_dtype(dtype: torch.dtype):
def get_model_architecture(
model_config: ModelConfig) -> Tuple[Type[nn.Module], str]:
architectures = getattr(model_config.hf_config, "architectures", [])
architectures = getattr(model_config.hf_config, "architectures", None) or []
# Special handling for quantized Mixtral.
# FIXME(woosuk): This is a temporary hack.
mixtral_supported = [