Revert "[NVIDIA]Fix local_num_experts for EP (#8779)" (#8797)

This commit is contained in:
Yineng Zhang
2025-08-04 23:30:43 -07:00
committed by GitHub
parent 873f384a51
commit 5e91fed1c5
2 changed files with 2 additions and 4 deletions

View File

@@ -200,8 +200,7 @@ class FusedMoE(torch.nn.Module):
self.quant_config = quant_config
self.quant_method.create_weights(
layer=self,
num_experts=self.num_experts,
num_local_experts=self.num_local_experts,
num_experts=self.num_local_experts,
hidden_size=hidden_size,
# FIXME: figure out which intermediate_size to use
intermediate_size=self.intermediate_size_per_partition,

View File

@@ -752,7 +752,6 @@ class ModelOptNvFp4FusedMoEMethod(FusedMoEMethodBase):
self,
layer: torch.nn.Module,
num_experts: int,
num_local_experts: int,
hidden_size: int,
intermediate_size_per_partition: int,
params_dtype: torch.dtype,
@@ -766,7 +765,7 @@ class ModelOptNvFp4FusedMoEMethod(FusedMoEMethodBase):
# TODO(ch-wan): check if this is needed
layer.num_experts = num_experts
layer.num_local_experts = num_local_experts
layer.num_local_experts = num_experts
layer.intermediate_size_per_partition = intermediate_size_per_partition
layer.params_dtype = params_dtype
layer.quant_config = self.quant_config