diff --git a/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py b/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py index 506bd29..6a41f9d 100644 --- a/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py +++ b/vllm-v0.6.2/vllm/model_executor/model_loader/utils.py @@ -20,7 +20,7 @@ def set_default_torch_dtype(dtype: torch.dtype): def get_model_architecture( model_config: ModelConfig) -> Tuple[Type[nn.Module], str]: - architectures = getattr(model_config.hf_config, "architectures", []) + architectures = getattr(model_config.hf_config, "architectures", None) or [] # Special handling for quantized Mixtral. # FIXME(woosuk): This is a temporary hack. mixtral_supported = [