2025-06-05 16:28:01 +08:00
# Additional Configuration
addintional configuration is a mechanism provided by vLLM to allow plugins to control inner behavior by their own. vLLM Ascend uses this mechanism to make the project more flexible.
## How to use
With either online mode or offline mode, users can use additional configuration. Take Qwen3 as an example:
**Online mode**:
```bash
vllm serve Qwen/Qwen3-8B --additional-config='{"config_key":"config_value"}'
```
**Offline mode**:
```python
from vllm import LLM
LLM(model="Qwen/Qwen3-8B", additional_config={"config_key":"config_value"})
```
### Configuration options
The following table lists the additional configuration options available in vLLM Ascend:
2025-06-09 19:28:11 +08:00
| Name | Type | Default | Description |
|-------------------------------| ---- |------|-----------------------------------------------------------------------------------------------|
| `torchair_graph_config` | dict | `{}` | The config options for torchair graph mode |
| `ascend_scheduler_config` | dict | `{}` | The config options for ascend scheduler |
| `expert_tensor_parallel_size` | str | `0` | Expert tensor parallel size the model to use. |
| `refresh` | bool | `false` | Whether to refresh global ascend config content. This value is usually used by rlhf case. |
| `expert_map_path` | str | None | When using expert load balancing for the MOE model, an expert map path needs to be passed in. |
2025-06-05 16:28:01 +08:00
The details of each config option are as follows:
**torchair_graph_config**
| Name | Type | Default | Description |
| ---- | ---- | ------- | ----------- |
| `enabled` | bool | `False` | Whether to enable torchair graph mode |
2025-06-07 14:28:20 +08:00
| `enable_view_optimize` | bool | `True` | Whether to enable torchair view optimization |
2025-06-05 16:28:01 +08:00
| `use_cached_graph` | bool | `False` | Whether to use cached graph |
| `graph_batch_sizes` | list[int] | `[]` | The batch size for torchair graph cache |
| `graph_batch_sizes_init` | bool | `False` | Init graph batch size dynamically if `graph_batch_sizes` is empty |
2025-06-06 18:54:37 +08:00
| `enable_multistream_shared_expert` | bool | `False` | Whether to enable multistream shared expert |
2025-06-05 16:28:01 +08:00
**ascend_scheduler_config**
| Name | Type | Default | Description |
| ---- | ---- | ------- | ----------- |
| `enabled` | bool | `False` | Whether to enable ascend scheduler for V1 engine|
ascend_scheduler_config also support the options from [vllm scheduler config ](https://docs.vllm.ai/en/stable/api/vllm/config.html#vllm.config.SchedulerConfig ). For example, you can add `chunked_prefill_enabled: true` to ascend_scheduler_config as well.
### Example
A full example of additional configuration is as follows:
```
{
"torchair_graph_config": {
"enabled": true,
"use_cached_graph": true,
"graph_batch_sizes": [1, 2, 4, 8],
2025-06-06 18:54:37 +08:00
"graph_batch_sizes_init": false,
"enable_multistream_shared_expert": false
2025-06-05 16:28:01 +08:00
},
"ascend_scheduler_config": {
"enabled": true,
"chunked_prefill_enabled": true,
},
2025-06-06 18:54:37 +08:00
"expert_tensor_parallel_size": 1,
"refresh": false,
2025-06-05 16:28:01 +08:00
}
```