From fff258bce17e05d787bfc5a225ed9d93db210f24 Mon Sep 17 00:00:00 2001 From: wangxiyuan Date: Fri, 21 Nov 2025 23:03:32 +0800 Subject: [PATCH] [Doc] add release note for v0.11.0rc2 (#4348) add release note for v0.11.0rc2 - vLLM version: v0.11.0 - vLLM main: https://github.com/vllm-project/vllm/commit/2918c1b49c88c29783c86f78d2c4221cb9622379 Signed-off-by: wangxiyuan --- README.md | 2 +- README.zh.md | 2 +- docs/source/community/versioning_policy.md | 2 ++ docs/source/conf.py | 6 ++--- docs/source/faqs.md | 2 +- docs/source/user_guide/release_notes.md | 28 ++++++++++++++++++++++ 6 files changed, 36 insertions(+), 6 deletions(-) diff --git a/README.md b/README.md index f67843be..0c3c27b1 100644 --- a/README.md +++ b/README.md @@ -52,7 +52,7 @@ Please use the following recommended versions to get started quickly: | Version | Release type | Doc | |------------|--------------|--------------------------------------| -|v0.11.0rc1|Latest release candidate|[QuickStart](https://vllm-ascend.readthedocs.io/en/latest/quick_start.html) and [Installation](https://vllm-ascend.readthedocs.io/en/latest/installation.html) for more details| +|v0.11.0rc2|Latest release candidate|[QuickStart](https://vllm-ascend.readthedocs.io/en/latest/quick_start.html) and [Installation](https://vllm-ascend.readthedocs.io/en/latest/installation.html) for more details| |v0.9.1|Latest stable version|[QuickStart](https://vllm-ascend.readthedocs.io/en/v0.9.1-dev/quick_start.html) and [Installation](https://vllm-ascend.readthedocs.io/en/v0.9.1-dev/installation.html) for more details| ## Contributing diff --git a/README.zh.md b/README.zh.md index bb88a6df..516c23a9 100644 --- a/README.zh.md +++ b/README.zh.md @@ -53,7 +53,7 @@ vLLM 昇腾插件 (`vllm-ascend`) 是一个由社区维护的让vLLM在Ascend NP | Version | Release type | Doc | |------------|--------------|--------------------------------------| -|v0.11.0rc1| 最新RC版本 |请查看[快速开始](https://vllm-ascend.readthedocs.io/en/latest/quick_start.html)和[安装指南](https://vllm-ascend.readthedocs.io/en/latest/installation.html)了解更多| +|v0.11.0rc2| 最新RC版本 |请查看[快速开始](https://vllm-ascend.readthedocs.io/en/latest/quick_start.html)和[安装指南](https://vllm-ascend.readthedocs.io/en/latest/installation.html)了解更多| |v0.9.1| 最新正式/稳定版本 |[快速开始](https://vllm-ascend.readthedocs.io/en/v0.9.1-dev/quick_start.html) and [安装指南](https://vllm-ascend.readthedocs.io/en/v0.9.1-dev/installation.html)了解更多| ## 贡献 diff --git a/docs/source/community/versioning_policy.md b/docs/source/community/versioning_policy.md index afc09212..f1044899 100644 --- a/docs/source/community/versioning_policy.md +++ b/docs/source/community/versioning_policy.md @@ -22,6 +22,7 @@ The table below is the release compatibility matrix for vLLM Ascend release. | vLLM Ascend | vLLM | Python | Stable CANN | PyTorch/torch_npu | MindIE Turbo | |-------------|--------------|------------------|-------------|--------------------|--------------| +| v0.11.0rc2 | v0.11.0 | >= 3.9, < 3.12 | 8.3.RC2 | 2.7.1 / 2.7.1 | | | v0.11.0rc1 | v0.11.0 | >= 3.9, < 3.12 | 8.3.RC1 | 2.7.1 / 2.7.1 | | | v0.11.0rc0 | v0.11.0rc3 | >= 3.9, < 3.12 | 8.2.RC1 | 2.7.1 / 2.7.1.dev20250724 | | | v0.10.2rc1 | v0.10.2 | >= 3.9, < 3.12 | 8.2.RC1 | 2.7.1 / 2.7.1.dev20250724 | | @@ -50,6 +51,7 @@ For main branch of vLLM Ascend, we usually make it compatible with the latest vL | Date | Event | |------------|-------------------------------------------| +| 2025.11.21 | Release candidates, v0.11.0rc2 | | 2025.11.10 | Release candidates, v0.11.0rc1 | | 2025.09.30 | Release candidates, v0.11.0rc0 | | 2025.09.16 | Release candidates, v0.10.2rc1 | diff --git a/docs/source/conf.py b/docs/source/conf.py index 4ca53ac0..0a2f7157 100644 --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -69,13 +69,13 @@ myst_substitutions = { # the branch of vllm-ascend, used in vllm-ascend clone and image tag # - main branch: 'main' # - vX.Y.Z branch: latest vllm-ascend release tag - 'vllm_ascend_version': 'v0.11.0rc1', + 'vllm_ascend_version': 'v0.11.0rc2', # the newest release version of vllm-ascend and matched vLLM, used in pip install. # This value should be updated when cut down release. - 'pip_vllm_ascend_version': "0.11.0rc1", + 'pip_vllm_ascend_version': "0.11.0rc2", 'pip_vllm_version': "0.11.0", # CANN image tag - 'cann_image_tag': "8.3.rc1-910b-ubuntu22.04-py3.11", + 'cann_image_tag': "8.3.rc2-910b-ubuntu22.04-py3.11", # vllm version in ci 'ci_vllm_version': 'v0.11.0', } diff --git a/docs/source/faqs.md b/docs/source/faqs.md index aea56b26..17caea8e 100644 --- a/docs/source/faqs.md +++ b/docs/source/faqs.md @@ -3,7 +3,7 @@ ## Version Specific FAQs - [[v0.9.1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/2643) -- [[v0.11.0rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/3222) +- [[v0.11.0rc2] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/4343) ## General FAQs diff --git a/docs/source/user_guide/release_notes.md b/docs/source/user_guide/release_notes.md index f1e2792c..4c0c0511 100644 --- a/docs/source/user_guide/release_notes.md +++ b/docs/source/user_guide/release_notes.md @@ -1,5 +1,33 @@ # Release Notes +## v0.11.0rc2 - 2025.11.21 +This is the second release candidate of v0.11.0 for vLLM Ascend. In this release, we solved many bugs to improve the quality. Thanks for all your feedback. We'll keep working on bug fix and performance improvement. The v0.11.0 official release will come soon. Please follow the [official doc](https://vllm-ascend.readthedocs.io/en/v0.11.0-dev) to get started. + +### Highlights +- CANN is upgraded to 8.3.RC2. [#4332](https://github.com/vllm-project/vllm-ascend/pull/4332) +- Ngram spec decode method is back now. [#4092](https://github.com/vllm-project/vllm-ascend/pull/4092) +- The performance of aclgraph is improved by updating default capture size. [#4205](https://github.com/vllm-project/vllm-ascend/pull/4205) + +### Core +- Speed up vLLM startup time. [#4099](https://github.com/vllm-project/vllm-ascend/pull/4099) +- Kimi k2 with quantization works now. [#4190](https://github.com/vllm-project/vllm-ascend/pull/4190) +- Fix a bug for qwen3-next. It's more stable now. [#4025](https://github.com/vllm-project/vllm-ascend/pull/4025) + +### Other +- Fix an issue for full decode only mode. Full graph mode is more stable now. [#4106](https://github.com/vllm-project/vllm-ascend/pull/4106) [#4282](https://github.com/vllm-project/vllm-ascend/pull/4282) +- Fix a allgather ops bug for DeepSeek V3 series models. [#3711](https://github.com/vllm-project/vllm-ascend/pull/3711) +- Fix some bugs for EPLB feature. [#4150](https://github.com/vllm-project/vllm-ascend/pull/4150) [#4334](https://github.com/vllm-project/vllm-ascend/pull/4334) +- Fix a bug that vl model doesn't work on x86 machine. [#4285](https://github.com/vllm-project/vllm-ascend/pull/4285) +- Support ipv6 for prefill disaggregation proxy. Please note that mooncake connector doesn't work with ipv6. We're working on it. [#4242](https://github.com/vllm-project/vllm-ascend/pull/4242) +- Add a check that to ensure EPLB only support w8a8 method for quantization case. [#4315](https://github.com/vllm-project/vllm-ascend/pull/4315) +- Add a check that to ensure FLASHCOMM feature doesn't work with vl model. It'll be supported in 2025 Q4 [#4222](https://github.com/vllm-project/vllm-ascend/pull/4222) +- Audio required library is installed in container. [#4324](https://github.com/vllm-project/vllm-ascend/pull/4324) + +### Known Issues +- Ray + EP doesn't work, if you run vLLM Ascend with ray, please disable expert parallelism. [#4123](https://github.com/vllm-project/vllm-ascend/pull/4123) +- `response_format` parameter is not supported yet. We'll support it soon. [#4175](https://github.com/vllm-project/vllm-ascend/pull/4175) +- cpu bind feature doesn't work for multi instance case(Such as multi DP on one node). We'll fix it in the next release. + ## v0.11.0rc1 - 2025.11.10 This is the first release candidate of v0.11.0 for vLLM Ascend. Please follow the [official doc](https://vllm-ascend.readthedocs.io/en/v0.11.0-dev) to get started.