From d9d1f3a7249c455a019f4fe0e7bb3b7c232ef392 Mon Sep 17 00:00:00 2001 From: Chranos <826995883@qq.com> Date: Wed, 11 Feb 2026 14:39:48 +0800 Subject: [PATCH] add deepseekv3 and llama4 --- vllm-v0.6.2/vllm/model_executor/model_loader/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py b/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py index 506bd29..6a41f9d 100644 --- a/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py +++ b/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py @@ -20,7 +20,7 @@ def set_default_torch_dtype(dtype: torch.dtype): def get_model_architecture( model_config: ModelConfig) -> Tuple[Type[nn.Module], str]: - architectures = getattr(model_config.hf_config, "architectures", []) + architectures = getattr(model_config.hf_config, "architectures", None) or [] # Special handling for quantized Mixtral. # FIXME(woosuk): This is a temporary hack. mixtral_supported = [