Remove fused_moe_grok (#2223)

This commit is contained in:
Lianmin Zheng
2024-11-27 14:28:55 -08:00
committed by GitHub
parent a4fd2f9b46
commit dd5eba4c88
7 changed files with 12 additions and 1372 deletions

View File

@@ -10,7 +10,7 @@ import triton.language as tl
from tqdm import tqdm
from transformers import AutoConfig
from sglang.srt.layers.fused_moe_grok.fused_moe import fused_moe, get_config_file_name
from sglang.srt.layers.fused_moe_triton.fused_moe import fused_moe, get_config_file_name
padding_size = 128 if bool(int(os.getenv("MOE_PADDING", "0"))) else 0