fix: resolve the fp8 bug introduced by vLLM 0.5.5 (#1276)

This commit is contained in:
Yineng Zhang
2024-09-01 00:44:29 +10:00
committed by GitHub
parent 51c554d812
commit 52cefdbf57

View File

@@ -695,7 +695,7 @@ class DeepseekV2ForCausalLM(nn.Module):
weight_loader( weight_loader(
param, param,
loaded_weight, loaded_weight,
weight_name, name,
shard_id=shard_id, shard_id=shard_id,
expert_id=expert_id, expert_id=expert_id,
) )