[quantization] Properly ignore quantization for layers excluded in quant_config (#11205)

This commit is contained in:
Bowen Bao
2025-10-07 14:06:05 -07:00
committed by GitHub
parent 420c99acfe
commit cd4b39a900
2 changed files with 8 additions and 10 deletions

View File

@@ -207,15 +207,11 @@ class FusedMoE(torch.nn.Module):
gemm1_clamp_limit=gemm1_clamp_limit,
)
if quant_config is None:
self.quant_method: FusedMoEMethodBase = UnquantizedFusedMoEMethod(
self.use_triton_kernels
)
else:
self.quant_method: FusedMoEMethodBase = quant_config.get_quant_method(
self, prefix
)
assert self.quant_method is not None
self.quant_method: Optional[FusedMoEMethodBase] = None
if quant_config is not None:
self.quant_method = quant_config.get_quant_method(self, prefix)
if self.quant_method is None:
self.quant_method = UnquantizedFusedMoEMethod(self.use_triton_kernels)
self.quant_method.create_weights(
layer=self,

View File

@@ -65,7 +65,9 @@ class QuarkConfig(QuantizationConfig):
if should_ignore_layer(
prefix, ignore=exclude_layers, fused_mapping=self.packed_modules_mapping
):
return UnquantizedLinearMethod()
if isinstance(layer, LinearBase):
return UnquantizedLinearMethod()
return None
if isinstance(layer, LinearBase):
scheme = self.get_scheme(layer=layer, layer_name=prefix)