Fix shared experts fusion error without quantization (#5632)
This commit is contained in:
@@ -1604,7 +1604,7 @@ class DeepseekV2ForCausalLM(nn.Module):
|
||||
if self.n_share_experts_fusion > 0:
|
||||
weights_list = list(weights)
|
||||
weights_dict = dict(weights_list)
|
||||
if self.quant_config.get_name() == "w8a8_int8":
|
||||
if self.quant_config is None or self.quant_config.get_name() == "w8a8_int8":
|
||||
suffix_list = [
|
||||
"down_proj.weight",
|
||||
"down_proj.weight_scale",
|
||||
|
||||
Reference in New Issue
Block a user