fix: resolve the fp8 bug introduced by vLLM 0.5.5 (#1276)

This commit is contained in:
Yineng Zhang
2024-09-01 00:44:29 +10:00
committed by GitHub
parent 51c554d812
commit 52cefdbf57

View File

@@ -695,7 +695,7 @@ class DeepseekV2ForCausalLM(nn.Module):
weight_loader(
param,
loaded_weight,
weight_name,
name,
shard_id=shard_id,
expert_id=expert_id,
)