[Doc] Update 0.9.0rc1 release date (#1139)
1. Update 0.9.0rc1 release date 2. Update feature and model support list 3. Add DP known issue to release note Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
This commit is contained in:
@@ -34,7 +34,7 @@ Following is the Release Compatibility Matrix for vLLM Ascend Plugin:
|
||||
|
||||
| Date | Event |
|
||||
|------------|-------------------------------------------|
|
||||
| 2025.06.07 | Release candidates, v0.9.0rc1 |
|
||||
| 2025.06.09 | Release candidates, v0.9.0rc1 |
|
||||
| 2025.05.29 | v0.7.x post release, v0.7.3.post1 |
|
||||
| 2025.05.08 | v0.7.x Final release, v0.7.3 |
|
||||
| 2025.05.06 | Release candidates, v0.8.5rc1 |
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
# Release note
|
||||
|
||||
## v0.9.0rc1 - 2025.06.07
|
||||
## v0.9.0rc1 - 2025.06.09
|
||||
|
||||
This is the 1st release candidate of v0.9.0 for vllm-ascend. Please follow the [official doc](https://vllm-ascend.readthedocs.io/en/) to start the journey. From this release, V1 Engine is recommended to use. The code of V0 Engine is frozen and will not be maintained any more. Please set environment `VLLM_USE_V1=1` to enable V1 Engine.
|
||||
|
||||
@@ -36,6 +36,7 @@ This is the 1st release candidate of v0.9.0 for vllm-ascend. Please follow the [
|
||||
### Known Issue
|
||||
|
||||
- In some case, vLLM process may be crashed with aclgraph enabled. We're working this issue and it'll be fixed in the next release.
|
||||
- Multi node data-parallel doesn't work with this release. This is a known issue in vllm and has been fixed on main branch. [#18981](https://github.com/vllm-project/vllm/pull/18981)
|
||||
|
||||
## v0.7.3.post1 - 2025.05.29
|
||||
|
||||
|
||||
@@ -18,13 +18,13 @@
|
||||
| Phi-4-mini | ✅ | |
|
||||
| MiniCPM | ✅ | |
|
||||
| MiniCPM3 | ✅ | |
|
||||
| LLama4 | ✅ | |
|
||||
| Mistral | | Need test |
|
||||
| DeepSeek v2.5 | |Need test |
|
||||
| Gemma-2 | | Need test |
|
||||
| Mllama | |Need test|
|
||||
| Gemma-3 | ❌ | [#496](https://github.com/vllm-project/vllm-ascend/issues/496) |
|
||||
| ChatGLM | ❌ | [#554](https://github.com/vllm-project/vllm-ascend/issues/554) |
|
||||
| LLama4 | ❌ | [#471](https://github.com/vllm-project/vllm-ascend/issues/471) |
|
||||
|
||||
### Pooling Models
|
||||
| Model | Supported | Note |
|
||||
|
||||
@@ -6,11 +6,11 @@ You can check the [support status of vLLM V1 Engine][v1_user_guide]. Below is th
|
||||
|
||||
| Feature | vLLM V0 Engine | vLLM V1 Engine | Next Step |
|
||||
|-------------------------------|----------------|----------------|------------------------------------------------------------------------|
|
||||
| Chunked Prefill | 🚧 WIP | 🟢 Functional | Functional, see detail note: [Chunked Prefill][cp] |
|
||||
| Automatic Prefix Caching | 🚧 WIP | 🟢 Functional | Functional, see detail note: [vllm-ascend#732][apc] |
|
||||
| Chunked Prefill | 🟢 Functional | 🟢 Functional | Functional, see detail note: [Chunked Prefill][cp] |
|
||||
| Automatic Prefix Caching | 🟢 Functional | 🟢 Functional | Functional, see detail note: [vllm-ascend#732][apc] |
|
||||
| LoRA | 🟢 Functional | 🟢 Functional | [vllm-ascend#396][multilora], [vllm-ascend#893][v1 multilora] |
|
||||
| Prompt adapter | 🔴 No plan | 🟡 Planned | Plan in 2025.06.30 |
|
||||
| Speculative decoding | 🟢 Functional | 🚧 WIP | CI needed; working on V1 support |
|
||||
| Prompt adapter | 🔴 No plan | 🔴 No plan | This feature has been deprecated by vllm. |
|
||||
| Speculative decoding | 🟢 Functional | 🟢 Functional | Basic support |
|
||||
| Pooling | 🟢 Functional | 🟡 Planned | CI needed and adapting more models; V1 support rely on vLLM support. |
|
||||
| Enc-dec | 🔴 NO plan | 🟡 Planned | Plan in 2025.06.30 |
|
||||
| Multi Modality | 🟢 Functional | 🟢 Functional | [Tutorial][multimodal], optimizing and adapting more models |
|
||||
|
||||
Reference in New Issue
Block a user