[Doc] Update user doc index (#1581)

Add user doc index to make the user guide more clear
- vLLM version: v0.9.1
- vLLM main:
49e8c7ea25

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
This commit is contained in:
wangxiyuan
2025-07-10 14:26:59 +08:00
committed by GitHub
parent c7446438a9
commit 3d1e6a5929
16 changed files with 42 additions and 28 deletions

View File

@@ -43,16 +43,10 @@ faqs
:::{toctree}
:caption: User Guide
:maxdepth: 1
user_guide/suppoted_features
user_guide/supported_models
user_guide/env_vars
user_guide/additional_config
user_guide/sleep_mode
user_guide/graph_mode.md
user_guide/lora.md
user_guide/quantization.md
user_guide/support_matrix/index
user_guide/configuration/index
user_guide/feature_guide/index
user_guide/release_notes
user_guide/structured_output
:::
% How to contribute to the vLLM Ascend project

View File

@@ -54,7 +54,7 @@ hccn_tool -i 0 -ping -g address 10.20.0.20
```
## Run with docker
Assume you have two Altas 800 A2(64G*8) nodes, and want to deploy the `deepseek-v3-w8a8` quantitative model across multi-node.
Assume you have two Atlas 800 A2(64G*8) nodes, and want to deploy the `deepseek-v3-w8a8` quantitative model across multi-node.
```shell
# Define the image and container name

View File

@@ -2,7 +2,7 @@
vllm-ascend uses the following environment variables to configure the system:
:::{literalinclude} ../../../vllm_ascend/envs.py
:::{literalinclude} ../../../../vllm_ascend/envs.py
:language: python
:start-after: begin-env-vars-definition
:end-before: end-env-vars-definition

View File

@@ -0,0 +1,10 @@
# Configuration Guide
This section provides a detailed configuration guide of vLLM Ascend.
:::{toctree}
:caption: Configuration Guide
:maxdepth: 1
env_vars
additional_config
:::

View File

@@ -59,7 +59,7 @@ online example:
vllm serve Qwen/Qwen2-7B-Instruct --additional-config='{"torchair_graph_config": {"enabled": true},"ascend_scheduler_config": {"enabled": true,}}'
```
You can find more detail about additional config [here](./additional_config.md)
You can find more detail about additional config [here](../configuration/additional_config.md).
## Fallback to Eager Mode

View File

Before

Width:  |  Height:  |  Size: 57 KiB

After

Width:  |  Height:  |  Size: 57 KiB

View File

@@ -0,0 +1,13 @@
# Feature Guide
This section provides a detailed usage guide of vLLM Ascend features.
:::{toctree}
:caption: Feature Guide
:maxdepth: 1
graph_mode
quantization
sleep_mode
structured_output
lora
:::

View File

@@ -1,4 +1,4 @@
# LoRA Adapters
# LoRA Adapters Guide
Like vLLM, vllm-ascend supports LoRA as well. The usage and more details can be found in [vLLM official document](https://docs.vllm.ai/en/latest/features/lora.html).

View File

@@ -1,4 +1,4 @@
# Sleep Mode
# Sleep Mode Guide
## Overview

View File

@@ -1,13 +0,0 @@
## {version}
### Highlights
- {feature}
### Bug fixes
- {bug}
### Other changes
- {change}
### Known issues
- {issue}
### Upgrade Notes
- {upgrade}
### Deprecation Notes
- {deprecation}

View File

@@ -0,0 +1,10 @@
# Features and models
This section provides a detailed supported matrix by vLLM Ascend.
:::{toctree}
:caption: Support Matrix
:maxdepth: 1
supported_models
supported_features
:::

View File

@@ -1,4 +1,4 @@
# Supported Models
# Model Support
## Text-only Language Models