fix: resolve the fp8 bug introduced by vLLM 0.5.5 (#1276)
This commit is contained in:
@@ -695,7 +695,7 @@ class DeepseekV2ForCausalLM(nn.Module):
|
|||||||
weight_loader(
|
weight_loader(
|
||||||
param,
|
param,
|
||||||
loaded_weight,
|
loaded_weight,
|
||||||
weight_name,
|
name,
|
||||||
shard_id=shard_id,
|
shard_id=shard_id,
|
||||||
expert_id=expert_id,
|
expert_id=expert_id,
|
||||||
)
|
)
|
||||||
|
|||||||
Reference in New Issue
Block a user