From 042605f4b2672828b6f9bc084307db58e7c4999f Mon Sep 17 00:00:00 2001 From: Li Wang Date: Wed, 27 Aug 2025 09:05:46 +0800 Subject: [PATCH] [Doc] Add stable modelslim branch (#2545) ### What this PR does / why we need it? The branch `br_release_MindStudio_8.1.RC2_TR5_20260624` is commercial delivery version of modelslim in Q3, and has been verified available ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? - vLLM version: v0.10.1.1 - vLLM main: https://github.com/vllm-project/vllm/commit/7d67a9d9f93f86b74066c64c373405aa088e4a16 Signed-off-by: wangli --- .../source/tutorials/single_npu_qwen3_quantization.md | 7 +++---- docs/source/user_guide/feature_guide/quantization.md | 11 +++++------ 2 files changed, 8 insertions(+), 10 deletions(-) diff --git a/docs/source/tutorials/single_npu_qwen3_quantization.md b/docs/source/tutorials/single_npu_qwen3_quantization.md index 56b4443..3092480 100644 --- a/docs/source/tutorials/single_npu_qwen3_quantization.md +++ b/docs/source/tutorials/single_npu_qwen3_quantization.md @@ -32,11 +32,10 @@ see https://www.modelscope.cn/models/vllm-ascend/Qwen3-8B-W4A8 ::: ```bash -git clone https://gitee.com/ascend/msit -cd msit/msmodelslim +# The branch(br_release_MindStudio_8.1.RC2_TR5_20260624) has been verified +git clone -b br_release_MindStudio_8.1.RC2_TR5_20260624 https://gitee.com/ascend/msit -# Optional, this commit has been verified -git checkout f8ab35a772a6c1ee7675368a2aa4bafba3bedd1a +cd msit/msmodelslim # Install by run this script bash install.sh diff --git a/docs/source/user_guide/feature_guide/quantization.md b/docs/source/user_guide/feature_guide/quantization.md index 1caa491..9e5f56c 100644 --- a/docs/source/user_guide/feature_guide/quantization.md +++ b/docs/source/user_guide/feature_guide/quantization.md @@ -11,11 +11,10 @@ To quantize a model, users should install [ModelSlim](https://gitee.com/ascend/m Install modelslim: ```bash -git clone https://gitee.com/ascend/msit -cd msit/msmodelslim +# The branch(br_release_MindStudio_8.1.RC2_TR5_20260624) has been verified +git clone -b br_release_MindStudio_8.1.RC2_TR5_20260624 https://gitee.com/ascend/msit -# Optional, this commit has been verified -git checkout f8ab35a772a6c1ee7675368a2aa4bafba3bedd1a +cd msit/msmodelslim bash install.sh pip install accelerate @@ -100,12 +99,12 @@ Enable quantization by specifying `--quantization ascend`, for more details, see ### 1. How to solve the KeyError: 'xxx.layers.0.self_attn.q_proj.weight' problem? -First, make sure you specify `ascend` quantization method. Second, check if your model is converted by this `modelslim-VLLM-8.1.RC1.b020_001` modelslim version. Finally, if it still doesn't work, please +First, make sure you specify `ascend` quantization method. Second, check if your model is converted by this `br_release_MindStudio_8.1.RC2_TR5_20260624` modelslim version. Finally, if it still doesn't work, please submit a issue, maybe some new models need to be adapted. ### 2. How to solve the error "Could not locate the configuration_deepseek.py"? -Please convert DeepSeek series models using `modelslim-VLLM-8.1.RC1.b020_001` modelslim, this version has fixed the missing configuration_deepseek.py error. +Please convert DeepSeek series models using `br_release_MindStudio_8.1.RC2_TR5_20260624` modelslim, this version has fixed the missing configuration_deepseek.py error. ### 3. When converting deepseek series models with modelslim, what should you pay attention?