Add activation parameters to fused_moe (#3170)

This commit is contained in:
Lianmin Zheng
2025-01-27 00:23:37 -08:00
committed by GitHub
parent 741fccd7bf
commit 52c03f16b9
7 changed files with 52 additions and 7 deletions

View File

@@ -133,6 +133,7 @@ class Grok1MoE(nn.Module):
renormalize=False,
quant_config=quant_config,
tp_size=tp_size,
activation="gelu",
use_presharded_weights=use_presharded_weights,
)