fix: resolve the fp8 bug introduced by vLLM 0.5.5 (#1276)
This commit is contained in:
@@ -695,7 +695,7 @@ class DeepseekV2ForCausalLM(nn.Module):
|
||||
weight_loader(
|
||||
param,
|
||||
loaded_weight,
|
||||
weight_name,
|
||||
name,
|
||||
shard_id=shard_id,
|
||||
expert_id=expert_id,
|
||||
)
|
||||
|
||||
Reference in New Issue
Block a user