From fe3be1595dc52ec3a10b784191b54b611804574f Mon Sep 17 00:00:00 2001 From: ylying <373472509@qq.com> Date: Wed, 17 Jul 2024 02:48:49 +0800 Subject: [PATCH] Add qwen2 tie word embedding (#630) --- python/sglang/srt/models/qwen2.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/python/sglang/srt/models/qwen2.py b/python/sglang/srt/models/qwen2.py index 83b7d2f9c..05152c271 100644 --- a/python/sglang/srt/models/qwen2.py +++ b/python/sglang/srt/models/qwen2.py @@ -313,6 +313,8 @@ class Qwen2ForCausalLM(nn.Module): param = params_dict[name] weight_loader = getattr(param, "weight_loader", default_weight_loader) weight_loader(param, loaded_weight) + if self.config.tie_word_embeddings and name=="model.embed_tokens.weight": + weight_loader(params_dict["lm_head.weight"], loaded_weight) EntryClass = Qwen2ForCausalLM