From ca8b1c3e47a4e51831128634f1aef4bbfdef2e6a Mon Sep 17 00:00:00 2001 From: wangxiyuan Date: Sat, 29 Mar 2025 09:02:08 +0800 Subject: [PATCH] [Doc] Add 0.7.3rc2 release note (#419) Add 0.7.3rc2 release note. We'll release 0.7.3rc2 right now. Signed-off-by: wangxiyuan --- .../developer_guide/versioning_policy.md | 11 ++++++---- docs/source/faqs.md | 1 + docs/source/installation.md | 4 ++-- docs/source/user_guide/release_notes.md | 22 +++++++++++++++++++ 4 files changed, 32 insertions(+), 6 deletions(-) diff --git a/docs/source/developer_guide/versioning_policy.md b/docs/source/developer_guide/versioning_policy.md index 976387f..72dbefe 100644 --- a/docs/source/developer_guide/versioning_policy.md +++ b/docs/source/developer_guide/versioning_policy.md @@ -67,15 +67,18 @@ Following is the Release Compatibility Matrix for vLLM Ascend Plugin: | vllm-ascend | vLLM | Python | Stable CANN | PyTorch/torch_npu | |--------------|--------------| --- | --- | --- | -| v0.7.3rc1 | v0.7.3 | 3.9 - 3.12 | 8.0.0 | 2.5.1 / 2.5.1.dev20250320 | +| v0.7.3rc2 | v0.7.3 | 3.9 - 3.12 | 8.0.0 | 2.5.1 / 2.5.1.dev20250320 | +| v0.7.3rc1 | v0.7.3 | 3.9 - 3.12 | 8.0.0 | 2.5.1 / 2.5.1.dev20250308 | | v0.7.1rc1 | v0.7.1 | 3.9 - 3.12 | 8.0.0 | 2.5.1 / 2.5.1.dev20250218 | ## Release cadence -### Next final release (`v0.7.3`) window +### release window | Date | Event | |------------|-------------------------------------------| +| 2025.04.15 | Release candidates, v0.8.Xrc1 | +| 2025.04.15 | Final release, v0.7.3 | +| 2025.03.28 | Release candidates, v0.7.3rc2 | | 2025.03.14 | Release candidates, v0.7.3rc1 | -| 2025.03.20 | Release candidates if needed, v0.7.3rc2 | -| 2025.03.30 | Final release, v0.7.3 | +| 2025.02.19 | Release candidates, v0.7.1rc1 | diff --git a/docs/source/faqs.md b/docs/source/faqs.md index 3f466b2..50067c4 100644 --- a/docs/source/faqs.md +++ b/docs/source/faqs.md @@ -4,6 +4,7 @@ - [[v0.7.1rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/19) - [[v0.7.3rc1] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/267) +- [[v0.7.3rc2] FAQ & Feedback](https://github.com/vllm-project/vllm-ascend/issues/418) ## General FAQs diff --git a/docs/source/installation.md b/docs/source/installation.md index 4b1ed1a..bec7eb2 100644 --- a/docs/source/installation.md +++ b/docs/source/installation.md @@ -158,9 +158,9 @@ Current version depends on a unreleased `torch-npu`, you need to install manuall # # Here we take python 3.10 on aarch64 as an example. Feel free to install the correct version for your environment. See: # -# https://pytorch-package.obs.cn-north-4.myhuaweicloud.com/pta/Daily/v2.5.1/20250308.3/pytorch_v2.5.1_py39.tar.gz +# https://pytorch-package.obs.cn-north-4.myhuaweicloud.com/pta/Daily/v2.5.1/20250320.3/pytorch_v2.5.1_py39.tar.gz # https://pytorch-package.obs.cn-north-4.myhuaweicloud.com/pta/Daily/v2.5.1/20250320.3/pytorch_v2.5.1_py310.tar.gz -# https://pytorch-package.obs.cn-north-4.myhuaweicloud.com/pta/Daily/v2.5.1/20250308.3/pytorch_v2.5.1_py311.tar.gz +# https://pytorch-package.obs.cn-north-4.myhuaweicloud.com/pta/Daily/v2.5.1/20250320.3/pytorch_v2.5.1_py311.tar.gz # mkdir pta cd pta diff --git a/docs/source/user_guide/release_notes.md b/docs/source/user_guide/release_notes.md index 1a60d47..fa04741 100644 --- a/docs/source/user_guide/release_notes.md +++ b/docs/source/user_guide/release_notes.md @@ -1,5 +1,27 @@ # Release note +## v0.7.3rc2 + +This is 2nd release candidate of v0.7.3 for vllm-ascend. Please follow the [official doc](https://vllm-ascend.readthedocs.io/en/v0.7.3-dev) to start the journey. +- Quickstart with container: https://vllm-ascend.readthedocs.io/en/v0.7.3-dev/quick_start.html +- Installation: https://vllm-ascend.readthedocs.io/en/v0.7.3-dev/installation.html + +### Highlights +- Add Ascend Custom Ops framewrok. Developers now can write customs ops using AscendC. An example ops `rotary_embedding` is added. More tutorials will come soon. The Custome Ops complation is disabled by default when installing vllm-ascend. Set `COMPILE_CUSTOM_KERNELS=1` to enable it. [#371](https://github.com/vllm-project/vllm-ascend/pull/371) +- V1 engine is basic supported in this release. The full support will be done in 0.8.X release. If you hit any issue or have any requirement of V1 engine. Please tell us [here](https://github.com/vllm-project/vllm-ascend/issues/414). [#376](https://github.com/vllm-project/vllm-ascend/pull/376) +- Prefix cache feature works now. You can set `enable_prefix_caching=True` to enable it. [#282](https://github.com/vllm-project/vllm-ascend/pull/282) + +### Core +- Bump torch_npu version to dev20250320.3 to improve accuracy to fix `!!!` output problem. [#406](https://github.com/vllm-project/vllm-ascend/pull/406) + +### Model +- The performance of Qwen2-vl is improved by optimizing patch embedding (Conv3D). [#398](https://github.com/vllm-project/vllm-ascend/pull/398) + +### Other + +- Fixed a bug to make sure multi step scheduler feature work. [#349](https://github.com/vllm-project/vllm-ascend/pull/349) +- Fixed a bug to make prefix cache feature works with correct accuracy. [#424](https://github.com/vllm-project/vllm-ascend/pull/424) + ## v0.7.3rc1 🎉 Hello, World! This is the first release candidate of v0.7.3 for vllm-ascend. Please follow the [official doc](https://vllm-ascend.readthedocs.io/en/v0.7.3-dev) to start the journey.