[CI]fix lint (#7641)

### What this PR does / why we need it?
This pull request addresses a linting issue by reordering a specific
configuration assignment within the `apply_config_platform_defaults`
method in `vllm_ascend/platform.py`. This change ensures compliance with
code style guidelines without altering the functional behavior of the
system.
### Does this PR introduce _any_ user-facing change?

### How was this patch tested?

- vLLM version: v0.18.0
- vLLM main:
ed359c497a

Signed-off-by: Wang Kunpeng <1289706727@qq.com>
This commit is contained in:
Wang Kunpeng
2026-03-25 18:48:10 +08:00
committed by GitHub
parent 05a561129e
commit ff1860bd81

View File

@@ -189,12 +189,6 @@ class NPUPlatform(Platform):
return min(max_num_seqs * decode_query_len, 512)
@classmethod
def apply_config_platform_defaults(cls, vllm_config: VllmConfig) -> None:
default_max_cg_capture_size = cls._get_default_max_cudagraph_capture_size(vllm_config)
if default_max_cg_capture_size is not None:
vllm_config.compilation_config.max_cudagraph_capture_size = default_max_cg_capture_size
@classmethod
def get_device_capability(cls, device_id: int = 0):
return None
@@ -209,6 +203,10 @@ class NPUPlatform(Platform):
pass_config.sp_min_token_num = get_sp_min_token_num(vllm_config)
logger.info(f"set sp_min_token_num to {pass_config.sp_min_token_num}")
default_max_cg_capture_size = cls._get_default_max_cudagraph_capture_size(vllm_config)
if default_max_cg_capture_size is not None:
vllm_config.compilation_config.max_cudagraph_capture_size = default_max_cg_capture_size
@classmethod
def get_device_name(cls, device_id: int = 0) -> str:
return torch.npu.get_device_name(device_id)