[CI]fix lint (#7641)
### What this PR does / why we need it?
This pull request addresses a linting issue by reordering a specific
configuration assignment within the `apply_config_platform_defaults`
method in `vllm_ascend/platform.py`. This change ensures compliance with
code style guidelines without altering the functional behavior of the
system.
### Does this PR introduce _any_ user-facing change?
### How was this patch tested?
- vLLM version: v0.18.0
- vLLM main:
ed359c497a
Signed-off-by: Wang Kunpeng <1289706727@qq.com>
This commit is contained in:
@@ -189,12 +189,6 @@ class NPUPlatform(Platform):
|
||||
|
||||
return min(max_num_seqs * decode_query_len, 512)
|
||||
|
||||
@classmethod
|
||||
def apply_config_platform_defaults(cls, vllm_config: VllmConfig) -> None:
|
||||
default_max_cg_capture_size = cls._get_default_max_cudagraph_capture_size(vllm_config)
|
||||
if default_max_cg_capture_size is not None:
|
||||
vllm_config.compilation_config.max_cudagraph_capture_size = default_max_cg_capture_size
|
||||
|
||||
@classmethod
|
||||
def get_device_capability(cls, device_id: int = 0):
|
||||
return None
|
||||
@@ -209,6 +203,10 @@ class NPUPlatform(Platform):
|
||||
pass_config.sp_min_token_num = get_sp_min_token_num(vllm_config)
|
||||
logger.info(f"set sp_min_token_num to {pass_config.sp_min_token_num}")
|
||||
|
||||
default_max_cg_capture_size = cls._get_default_max_cudagraph_capture_size(vllm_config)
|
||||
if default_max_cg_capture_size is not None:
|
||||
vllm_config.compilation_config.max_cudagraph_capture_size = default_max_cg_capture_size
|
||||
|
||||
@classmethod
|
||||
def get_device_name(cls, device_id: int = 0) -> str:
|
||||
return torch.npu.get_device_name(device_id)
|
||||
|
||||
Reference in New Issue
Block a user