fixed lm_head.weight error for quantized qwen (#2910)
This commit is contained in:
@@ -356,6 +356,8 @@ class Qwen2ForCausalLM(nn.Module):
|
|||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
# Skip loading extra bias for GPTQ models.
|
# Skip loading extra bias for GPTQ models.
|
||||||
|
if "lm_head.weight" in name:
|
||||||
|
continue
|
||||||
if name.endswith(".bias") and name not in params_dict:
|
if name.endswith(".bias") and name not in params_dict:
|
||||||
continue
|
continue
|
||||||
param = params_dict[name]
|
param = params_dict[name]
|
||||||
|
|||||||
Reference in New Issue
Block a user