[Fix] Compatibility between DP attention and pipeline parallelism (#10100)
Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
This commit is contained in:
@@ -32,6 +32,7 @@ from sglang.srt.configs.model_config import AttentionArch, ModelConfig
|
||||
from sglang.srt.configs.update_config import adjust_config_with_unaligned_cpu_tp
|
||||
from sglang.srt.constants import GPU_MEMORY_TYPE_WEIGHTS
|
||||
from sglang.srt.distributed import (
|
||||
get_pp_group,
|
||||
get_tp_group,
|
||||
get_world_group,
|
||||
init_distributed_environment,
|
||||
@@ -639,6 +640,7 @@ class ModelRunner:
|
||||
cpu_group=get_world_group().cpu_group,
|
||||
)
|
||||
self.tp_group = get_tp_group()
|
||||
self.pp_group = get_pp_group()
|
||||
self.attention_tp_group = get_attention_tp_group()
|
||||
|
||||
# Check memory for tensor parallelism
|
||||
@@ -1825,7 +1827,10 @@ class ModelRunner:
|
||||
else:
|
||||
raise ValueError(f"Invalid forward mode: {forward_batch.forward_mode}")
|
||||
|
||||
if forward_batch.global_num_tokens_cpu is not None:
|
||||
if (
|
||||
forward_batch.global_num_tokens_cpu is not None
|
||||
and self.pp_group.is_last_rank
|
||||
):
|
||||
forward_batch.post_forward_mlp_sync_batch(ret)
|
||||
|
||||
return ret, can_run_cuda_graph
|
||||
|
||||
Reference in New Issue
Block a user