[Bugs] Fix moe when without bias (#76)
This commit is contained in:
@@ -196,7 +196,9 @@ class FusedMoE(VllmFusedMoE):
|
|||||||
self.moe_config = moe
|
self.moe_config = moe
|
||||||
self.quant_config = quant_config
|
self.quant_config = quant_config
|
||||||
self.has_bias=has_bias
|
self.has_bias=has_bias
|
||||||
|
self.register_parameter("w13_bias", None)
|
||||||
|
self.register_parameter("w2_bias", None)
|
||||||
|
|
||||||
# Note: get_quant_method will look at the layer's local_num_experts
|
# Note: get_quant_method will look at the layer's local_num_experts
|
||||||
# for heuristic purposes, so it must be initialized first.
|
# for heuristic purposes, so it must be initialized first.
|
||||||
quant_method: Optional[QuantizeMethodBase] = None
|
quant_method: Optional[QuantizeMethodBase] = None
|
||||||
|
|||||||
Reference in New Issue
Block a user