SILONG ZENG
7faa6878a6
[Lint]Style: Convert vllm-ascend/ to ruff format(Batch #3 ) ( #5978 )
...
### What this PR does / why we need it?
**Scope of Changes**:
| File Path |
| :--- |
| `vllm_ascend/attention/mla_v1.py` |
| `vllm_ascend/attention/sfa_v1.py` |
| `vllm_ascend/core/recompute_scheduler.py` |
| `vllm_ascend/core/scheduler_dynamic_batch.py` |
| `vllm_ascend/distributed/device_communicators/npu_communicator.py` |
| `vllm_ascend/distributed/device_communicators/pyhccl.py` |
| `vllm_ascend/distributed/device_communicators/pyhccl_wrapper.py` |
### Does this PR introduce _any_ user-facing change?
### How was this patch tested?
- vLLM version: v0.13.0
- vLLM main:
2c24bc6996
---------
Signed-off-by: MrZ20 <2609716663@qq.com >
Co-authored-by: Soren <user@SorendeMac-mini.local >
2026-01-24 22:10:18 +08:00
lty
295018ec0f
[Refactor]Refactor of vllm_ascend/distributed module ( #5719 )
...
### What this PR does / why we need it?
Based on the RFC:https://github.com/vllm-project/vllm-ascend/issues/5604
This PR is a refactoring of vllm_ascend/distributed, moving all
kv_transfer realtaed codes into a dedicated folder, which has already
been done in vLLM
### Does this PR introduce _any_ user-facing change?
NA
### How was this patch tested?
- vLLM version: v0.13.0
- vLLM main:
2f4e6548ef
---------
Signed-off-by: lty <linhebiwen@gmail.com >
2026-01-15 08:57:40 +08:00
wangxiyuan
b917361ca5
[MISC] Clean up torch_npu ( #688 )
...
torch_npu 2.5.1 support autoload now. This patch does:
1. remove useless torch_npu import
2. replace `torch_npu.npu` to `torch.npu`.
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com >
2025-04-29 18:03:38 +08:00
Huazhong Ji
c3d1a3782a
Add pyhccl ( #503 )
...
This is the first step to support trl vllm serve on Ascend NPU
https://github.com/vllm-project/vllm-ascend/issues/459 .
This PR can work properly only when
https://github.com/vllm-project/vllm/pull/16464 is merged into vLLM.
---------
Signed-off-by: hzji210@gmail.com <hzji210@gmail.com >
2025-04-17 14:57:52 +08:00