From 40dee08f7b9938dbb35ef856c810a86cac0b7e3f Mon Sep 17 00:00:00 2001 From: Chranos <826995883@qq.com> Date: Mon, 9 Feb 2026 17:47:18 +0800 Subject: [PATCH] fix: handle missing tie_word_embeddings attr in MPTConfig Use getattr with default True for MPTConfig.tie_word_embeddings, as some MPT model configs lack this attribute. Co-Authored-By: Claude Opus 4.6 (1M context) --- vllm-v0.6.2/vllm/model_executor/models/mpt.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm-v0.6.2/vllm/model_executor/models/mpt.py b/vllm-v0.6.2/vllm/model_executor/models/mpt.py index e15c0fe..cfa020d 100644 --- a/vllm-v0.6.2/vllm/model_executor/models/mpt.py +++ b/vllm-v0.6.2/vllm/model_executor/models/mpt.py @@ -272,7 +272,7 @@ class MPTForCausalLM(nn.Module, SupportsPP): config = vllm_config.model_config.hf_config quant_config = vllm_config.quant_config self.config = config - assert config.tie_word_embeddings + assert getattr(config, "tie_word_embeddings", True) self.quant_config = quant_config self.transformer = MPTModel(vllm_config=vllm_config,