From 9254a33ad46d226069dc6a60e7c2e40abde96920 Mon Sep 17 00:00:00 2001 From: Xiaoyu Zhang <35585791+BBuf@users.noreply.github.com> Date: Sat, 28 Dec 2024 14:01:35 +0800 Subject: [PATCH] avoid fused_moe_triton `padding` circular import (#2624) --- python/sglang/srt/layers/quantization/fp8.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/python/sglang/srt/layers/quantization/fp8.py b/python/sglang/srt/layers/quantization/fp8.py index 989e5cd5c..bf8c58e49 100644 --- a/python/sglang/srt/layers/quantization/fp8.py +++ b/python/sglang/srt/layers/quantization/fp8.py @@ -28,7 +28,6 @@ from vllm.model_executor.layers.quantization.utils.w8a8_utils import ( from vllm.model_executor.parameter import ModelWeightParameter, PerTensorScaleParameter from sglang.srt.layers.linear import LinearMethodBase, UnquantizedLinearMethod -from sglang.srt.layers.moe.fused_moe_triton.fused_moe import padding_size from sglang.srt.layers.quantization.base_config import ( QuantizationConfig, QuantizeMethodBase, @@ -548,6 +547,10 @@ class Fp8MoEMethod: layer.w2_input_scale = None def process_weights_after_loading(self, layer: Module) -> None: + from sglang.srt.layers.moe.fused_moe_triton.fused_moe import ( + padding_size, # Avoid circular import + ) + # Block quant doesn't need to process weights after loading if self.block_quant: return