diff --git a/vllm-v0.6.2/vllm/model_executor/models/registry.py b/vllm-v0.6.2/vllm/model_executor/models/registry.py index 276fba1..dde8f02 100644 --- a/vllm-v0.6.2/vllm/model_executor/models/registry.py +++ b/vllm-v0.6.2/vllm/model_executor/models/registry.py @@ -483,6 +483,7 @@ class _ModelRegistry: architectures = [architectures] if not architectures: logger.warning("No model architectures are specified") + return [] return architectures diff --git a/vllm-v0.6.2/vllm/transformers_utils/config.py b/vllm-v0.6.2/vllm/transformers_utils/config.py index 4cd7e1f..c5b9daf 100644 --- a/vllm-v0.6.2/vllm/transformers_utils/config.py +++ b/vllm-v0.6.2/vllm/transformers_utils/config.py @@ -229,6 +229,14 @@ def get_config( model_type = MODEL_FOR_CAUSAL_LM_MAPPING_NAMES[config.model_type] config.update({"architectures": [model_type]}) + # Some composite config classes (e.g. Llama4Config) may not preserve + # the 'architectures' field from config.json. Restore it from the + # raw config_dict if needed. + if config_format == ConfigFormat.HF: + raw_archs = config_dict.get("architectures") + if raw_archs and not getattr(config, "architectures", None): + config.architectures = raw_archs + patch_rope_scaling(config) if trust_remote_code: