Quick Fix GLM (#9264)

This commit is contained in:
Stefan He
2025-08-16 23:43:41 -07:00
committed by GitHub
parent bb10e3a1c3
commit e47800e176
2 changed files with 6 additions and 1 deletions

View File

@@ -24,6 +24,7 @@ from transformers import PretrainedConfig
from sglang.srt.distributed import (
get_moe_expert_parallel_world_size,
get_pp_group,
get_tensor_model_parallel_rank,
get_tensor_model_parallel_world_size,
parallel_state,
@@ -719,6 +720,9 @@ class Glm4MoeModel(DeepseekV2Model):
for layer_id in range(config.num_hidden_layers)
]
)
self.pp_group = get_pp_group()
self.start_layer = 0
self.end_layer = config.num_hidden_layers
self.norm = RMSNorm(config.hidden_size, eps=config.rms_norm_eps)
@@ -735,6 +739,7 @@ class Glm4MoeForCausalLM(DeepseekV2ForCausalLM):
self.config = config
self.tp_size = get_tensor_model_parallel_world_size()
self.quant_config = quant_config
self.pp_group = get_pp_group()
self.determine_num_fused_shared_experts("Glm4MoeForCausalLM")
self.model = Glm4MoeModel(
config, quant_config, prefix=add_prefix("model", prefix)