[Doc] Update feature support list (#650)
1. remove Chinese doc. The content is out of data and we don't have enough time to maintain it. 2. Update feature support matrix. Refresh the content and add V1 status. --------- Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com> Signed-off-by: Yikun Jiang <yikunkero@gmail.com> Co-authored-by: Yikun Jiang <yikunkero@gmail.com>
This commit is contained in:
@@ -1,21 +1,44 @@
|
||||
# Feature Support
|
||||
|
||||
| Feature | Supported | CI Coverage | Guidance Document | Current Status | Next Step |
|
||||
|--------------------------|-----------|-------------|-------------------|---------------------------|--------------------|
|
||||
| Chunked Prefill | ❌ | | | NA | Plan in 2025.03.30 |
|
||||
| Automatic Prefix Caching | ❌ | | | NA | Plan in 2025.03.30 |
|
||||
| LoRA | ❌ | | | NA | Plan in 2025.06.30 |
|
||||
| Prompt adapter | ❌ | | | NA | Plan in 2025.06.30 |
|
||||
| Speculative decoding | ✅ | | | Basic functions available | Need fully test |
|
||||
| Pooling | ✅ | | | Basic functions available(Bert) | Need fully test and add more models support|
|
||||
| Enc-dec | ❌ | | | NA | Plan in 2025.06.30|
|
||||
| Multi Modality | ✅ | | ✅ | Basic functions available(LLaVA/Qwen2-vl/Qwen2-audio/internVL)| Improve performance, and add more models support |
|
||||
| LogProbs | ✅ | | | Basic functions available | Need fully test |
|
||||
| Prompt logProbs | ✅ | | | Basic functions available | Need fully test |
|
||||
| Async output | ✅ | | | Basic functions available | Need fully test |
|
||||
| Multi step scheduler | ✅ | | | Basic functions available | Need fully test, Find more details at [<u> Blog </u>](https://blog.vllm.ai/2024/09/05/perf-update.html#batch-scheduling-multiple-steps-ahead-pr-7000), [<u> RFC </u>](https://github.com/vllm-project/vllm/issues/6854) and [<u>issue</u>](https://github.com/vllm-project/vllm/pull/7000) |
|
||||
| Best of | ✅ | | | Basic functions available | Need fully test |
|
||||
| Beam search | ✅ | | | Basic functions available | Need fully test |
|
||||
| Guided Decoding | ✅ | | | Basic functions available | Find more details at the [<u>issue</u>](https://github.com/vllm-project/vllm-ascend/issues/177) |
|
||||
| Tensor Parallel | ✅ | | | Basic functions available | Need fully test |
|
||||
| Pipeline Parallel | ✅ | | | Basic functions available | Need fully test |
|
||||
The feature support principle of vLLM Ascend is: **aligned with the vLLM**. We are also actively collaborating with the community to accelerate support.
|
||||
|
||||
You can check the [support status of vLLM V1 Engine][v1_user_guide]. Below is the feature support status of vLLM Ascend:
|
||||
|
||||
| Feature | vLLM V0 Engine | vLLM V1 Engine | Next Step |
|
||||
|-------------------------------|----------------|----------------|------------------------------------------------------------------------|
|
||||
| Chunked Prefill | 🚧 WIP | 🚧 WIP | Functional, waiting for CANN 8.1 nnal package release |
|
||||
| Automatic Prefix Caching | 🚧 WIP | 🚧 WIP | Functional, waiting for CANN 8.1 nnal package release |
|
||||
| LoRA | 🟢 Functional | 🚧 WIP | [vllm-ascend#396][multilora], CI needed, working on V1 support |
|
||||
| Prompt adapter | No plan | 🟡 Planned | Plan in 2025.06.30 |
|
||||
| Speculative decoding | 🟢 Functional | 🚧 WIP | CI needed; working on V1 support |
|
||||
| Pooling | 🟢 Functional | 🟢 Functional | CI needed and adapting more models; V1 support rely on vLLM support. |
|
||||
| Enc-dec | 🔴 NO plan | 🟡 Planned | Plan in 2025.06.30 |
|
||||
| Multi Modality | 🟢 Functional | 🟢 Functional | [Tutorial][multimodal], optimizing and adapting more models |
|
||||
| LogProbs | 🟢 Functional | 🟢 Functional | CI needed |
|
||||
| Prompt logProbs | 🟢 Functional | 🟢 Functional | CI needed |
|
||||
| Async output | 🟢 Functional | 🟢 Functional | CI needed |
|
||||
| Multi step scheduler | 🟢 Functional | 🔴 Deprecated | [vllm#8779][v1_rfc], replaced by [vLLM V1 Scheduler][v1_scheduler]) |
|
||||
| Best of | 🟢 Functional | 🔴 Deprecated | [vllm#13361][best_of], CI needed |
|
||||
| Beam search | 🟢 Functional | 🟢 Functional | CI needed |
|
||||
| Guided Decoding | 🟢 Functional | 🟢 Functional | [vllm-ascend#177][guided_decoding] |
|
||||
| Tensor Parallel | 🟢 Functional | 🟢 Functional | CI needed |
|
||||
| Pipeline Parallel | 🟢 Functional | 🟢 Functional | CI needed |
|
||||
| Expert Parallel | 🔴 NO plan | 🟢 Functional | CI needed; No plan on V0 support |
|
||||
| Data Parallel | 🔴 NO plan | 🟢 Functional | CI needed; No plan on V0 support |
|
||||
| Prefill Decode Disaggregation | 🟢 Functional | 🟢 Functional | 1P1D available, working on xPyD and V1 support. |
|
||||
| Quantization | 🟢 Functional | 🟢 Functional | W8A8 available, CI needed; working on more quantization method support |
|
||||
| Graph Mode | 🔴 NO plan | 🟢 Functional | Functional, waiting for CANN 8.1 nnal package release |
|
||||
| Sleep Mode | 🟢 Functional | 🟢 Functional | level=1 available, CI needed, working on V1 support |
|
||||
|
||||
- 🟢 Functional: Fully operational, with ongoing optimizations.
|
||||
- 🚧 WIP: Under active development
|
||||
- 🟡 Planned: Scheduled for future implementation (some may have open PRs/RFCs).
|
||||
- 🔴 NO plan / Deprecated: No plan for V0 or deprecated by vLLM v1.
|
||||
|
||||
[v1_user_guide]: https://docs.vllm.ai/en/latest/getting_started/v1_user_guide.html
|
||||
[multimodal]: https://vllm-ascend.readthedocs.io/en/latest/tutorials/single_npu_multimodal.html
|
||||
[best_of]: https://github.com/vllm-project/vllm/issues/13361
|
||||
[guided_decoding]: https://github.com/vllm-project/vllm-ascend/issues/177
|
||||
[v1_scheduler]: https://github.com/vllm-project/vllm/blob/main/vllm/v1/core/sched/scheduler.py
|
||||
[v1_rfc]: https://github.com/vllm-project/vllm/issues/8779
|
||||
[multilora]: https://github.com/vllm-project/vllm-ascend/issues/396
|
||||
|
||||
Reference in New Issue
Block a user