From 4d921f2b7916c4ccc660a26cda4fd3e47f3d8536 Mon Sep 17 00:00:00 2001 From: Cheng Wan <54331508+ch-wan@users.noreply.github.com> Date: Sun, 27 Jul 2025 01:24:10 -0700 Subject: [PATCH] [hotfix] fix merge conflicts in FlashInferEPMoE (#8405) --- python/sglang/srt/layers/moe/ep_moe/layer.py | 1 + 1 file changed, 1 insertion(+) diff --git a/python/sglang/srt/layers/moe/ep_moe/layer.py b/python/sglang/srt/layers/moe/ep_moe/layer.py index e99946869..6aa83dc00 100644 --- a/python/sglang/srt/layers/moe/ep_moe/layer.py +++ b/python/sglang/srt/layers/moe/ep_moe/layer.py @@ -1236,6 +1236,7 @@ class FlashInferEPMoE(EPMoE): self.num_expert_group = num_expert_group self.topk_group = topk_group self.correction_bias = correction_bias + self.use_flashinfer_trtllm_moe = use_flashinfer_trtllm_moe def forward(self, hidden_states: torch.Tensor, router_logits: torch.Tensor): assert use_flashinfer_trtllm_moe