Feature: Fix the binding error in Llama (#3355)

This commit is contained in:
Chayenne
2025-02-06 20:19:24 -08:00
committed by GitHub
parent 823148e7f0
commit 40022d075a

View File

@@ -476,9 +476,14 @@ class LlamaForCausalLM(nn.Module):
# Skip loading kv_scale from ckpts towards new design.
if name.endswith(".kv_scale") and name not in params_dict:
continue
param = params_dict[name]
weight_loader = getattr(param, "weight_loader", default_weight_loader)
weight_loader(param, loaded_weight)
if name in params_dict.keys():
param = params_dict[name]
weight_loader = getattr(
param, "weight_loader", default_weight_loader
)
weight_loader(param, loaded_weight)
else:
logger.warning(f"Parameter {name} not found in params_dict")
def get_weights_by_name(
self, name: str, truncate_size: int = 100, tp_size: int = 1