[bugfix] fix typo of _skip_all_reduce_across_dp_group (#5435)
### What this PR does / why we need it?
fix typo of _skip_all_reduce_across_dp_group
### Does this PR introduce _any_ user-facing change?
no
### How was this patch tested?
- vLLM version: release/v0.13.0
- vLLM main:
81786c8774
Signed-off-by: jiangkuaixue123 <jiangxiaozhou111@163.com>
This commit is contained in:
@@ -397,7 +397,7 @@ class NPUModelRunner(GPUModelRunner):
|
||||
def _use_aclgraph(self) -> bool:
|
||||
return self.compilation_config.cudagraph_mode != CUDAGraphMode.NONE and self.compilation_config.mode == CompilationMode.VLLM_COMPILE and not self.model_config.enforce_eager
|
||||
|
||||
def _skip_all_reduce_acorss_dp_group(self) -> bool:
|
||||
def _skip_all_reduce_across_dp_group(self) -> bool:
|
||||
"""
|
||||
Decide whether to skip the all-reduce across the data-parallel (DP) group.
|
||||
|
||||
@@ -446,7 +446,7 @@ class NPUModelRunner(GPUModelRunner):
|
||||
if self.dp_size == 1:
|
||||
return num_tokens, None, with_prefill
|
||||
|
||||
if self._skip_all_reduce_acorss_dp_group():
|
||||
if self._skip_all_reduce_across_dp_group():
|
||||
num_tokens_after_padding = torch.tensor([num_tokens] *
|
||||
self.dp_size,
|
||||
device="cpu",
|
||||
|
||||
Reference in New Issue
Block a user