From c9b5881bcd3a7b76d7b631471e4f3eae2ac545dd Mon Sep 17 00:00:00 2001 From: "rongfu.leng" Date: Tue, 23 Dec 2025 10:21:46 +0800 Subject: [PATCH] [Doc] fix docs set rope_theta value is 10e6 in qwen3-235b model (#5258) ### What this PR does / why we need it? Fixes https://github.com/vllm-project/vllm-ascend/issues/5201 ### Does this PR introduce _any_ user-facing change? No, doc only ### How was this patch tested? - vLLM version: release/v0.13.0 - vLLM main: https://github.com/vllm-project/vllm/commit/ad32e3e19ccf0526cb6744a5fed09a138a5fb2f9 Signed-off-by: rongfu.leng --- docs/source/tutorials/Qwen3-235B-A22B.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/tutorials/Qwen3-235B-A22B.md b/docs/source/tutorials/Qwen3-235B-A22B.md index 227e0ce4..38b23111 100644 --- a/docs/source/tutorials/Qwen3-235B-A22B.md +++ b/docs/source/tutorials/Qwen3-235B-A22B.md @@ -126,7 +126,7 @@ vllm serve vllm-ascend/Qwen3-235B-A22B-w8a8 \ **Notice:** - for vllm version below `v0.12.0` use parameter: `--rope_scaling '{"rope_type":"yarn","factor":4,"original_max_position_embeddings":32768}' \` -- for vllm version `v0.12.0` use parameter: `--hf-overrides '{"rope_parameters": {"rope_type":"yarn","rope_theta":1000,"factor":4,"original_max_position_embeddings":32768}}' \` +- for vllm version `v0.12.0` use parameter: `--hf-overrides '{"rope_parameters": {"rope_type":"yarn","rope_theta":1000000,"factor":4,"original_max_position_embeddings":32768}}' \` The parameters are explained as follows: - `--data-parallel-size` 1 and `--tensor-parallel-size` 8 are common settings for data parallelism (DP) and tensor parallelism (TP) sizes.