[doc]Table split (#5929)

### What this PR does / why we need it?
Added legend descriptions, and split redundant tables into core
supported model tables and extended compatible model tables.
### Does this PR introduce _any_ user-facing change?
no

### How was this patch tested?
ut
- vLLM version: v0.13.0
- vLLM main:
11b6af5280

---------

Signed-off-by: herizhen <1270637059@qq.com>
This commit is contained in:
herizhen
2026-01-19 09:15:04 +08:00
committed by GitHub
parent c4fde5c064
commit 0eafed9bd6

View File

@@ -2,85 +2,105 @@
Get the latest info here: <https://github.com/vllm-project/vllm-ascend/issues/1608>
**Legend Description**:
- ✅ = Supported model/feature
- ❌ = Not supported model/feature
- 🟡 = Not tested or verified
## Text-Only Language Models
### Generative Models
#### Core Supported Models
| Model | Support | Note | BF16 | Supported Hardware | W8A8 | Chunked Prefill | Automatic Prefix Cache | LoRA | Speculative Decoding | Async Scheduling | Tensor Parallel | Pipeline Parallel | Expert Parallel | Data Parallel | Prefill-decode Disaggregation | Piecewise AclGraph | Fullgraph AclGraph | max-model-len | MLP Weight Prefetch | Doc |
|-------------------------------|-----------|----------------------------------------------------------------------|------|--------------------|------|-----------------|------------------------|------|----------------------|------------------|-----------------|-------------------|-----------------|---------------|-------------------------------|--------------------|--------------------|---------------|---------------------|-----|
| DeepSeek V3/3.1 | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ || ✅ || ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | 240k || [DeepSeek-V3.1](../../tutorials/DeepSeek-V3.1.md) |
| DeepSeek V3.2 | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | 160k | ✅ | [DeepSeek-V3.2](../../tutorials/DeepSeek-V3.2.md) |
| DeepSeek R1 | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ || ✅ || ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | 128k || [DeepSeek-R1](../../tutorials/DeepSeek-R1.md) |
| DeepSeek Distill (Qwen/Llama) | ✅ | || A2/A3 |||||||||||||||||
| Qwen3 | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ ||| ✅ | ✅ ||| ✅ || ✅ | ✅ | 128k | ✅ | [Qwen3-Dense](../../tutorials/Qwen3-Dense.md) |
| Qwen3-based | ✅ | || A2/A3 |||||||||||||||||
| Qwen3-Coder | ✅ | | ✅ | A2/A3 ||✅|✅|✅|||✅|✅|✅|✅||||||[Qwen3-Coder-30B-A3B tutorial](../../tutorials/Qwen3-Coder-30B-A3B.md)|
| Qwen3-Moe | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ ||| ✅ | ✅ || ✅ | ✅ | ✅ | ✅ | ✅ | 256k || [Qwen3-235B-A22B](../../tutorials/Qwen3-235B-A22B.md) |
| Qwen3-Next | ✅ | | ✅ | A2/A3 | ✅ |||||| ✅ ||| ✅ || ✅ | ✅ ||| [Qwen3-Next](../../tutorials/Qwen3-Next.md) |
| Qwen2.5 | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ |||| ✅ ||| ✅ |||||| [Qwen2.5-7B](../../tutorials/Qwen2.5-7B.md) |
| Qwen2 | ✅ | || A2/A3 |||||||||||||||||
| Qwen2-based | ✅ | || A2/A3 |||||||||||||||||
| QwQ-32B | ✅ | || A2/A3 |||||||||||||||||
| Llama2/3/3.1/3.2 | ✅ | || A2/A3 |||||||||||||||||
| Internlm | ✅ | [#1962](https://github.com/vllm-project/vllm-ascend/issues/1962) || A2/A3 |||||||||||||||||
| Baichuan | ✅ | || A2/A3 |||||||||||||||||
| Baichuan2 | ✅ | || A2/A3 |||||||||||||||||
| Phi-4-mini | ✅ | || A2/A3 |||||||||||||||||
| MiniCPM | ✅ | || A2/A3 |||||||||||||||||
| MiniCPM3 | ✅ | || A2/A3 |||||||||||||||||
| Ernie4.5 | ✅ | || A2/A3 |||||||||||||||||
| Ernie4.5-Moe | ✅ | || A2/A3 |||||||||||||||||
| Gemma-2 | ✅ | || A2/A3 |||||||||||||||||
| Gemma-3 | ✅ | || A2/A3 |||||||||||||||||
| Phi-3/4 | ✅ | || A2/A3 |||||||||||||||||
| Mistral/Mistral-Instruct | ✅ | || A2/A3 |||||||||||||||||
| GLM-4.x | ✅ | || A2/A3 |✅|✅|✅||✅|✅|✅|||✅||✅|✅|128k||../../tutorials/GLM4.x.md|
| GLM-4.x | ✅ | || A2/A3 |✅|✅|✅||✅|✅|✅|||||✅|✅|128k||[GLM-4.x](../../tutorials/GLM4.x.md)|
| Kimi-K2-Thinking | ✅ | || A2/A3 |||||||||||||||| [Kimi-K2-Thinking](../../tutorials/Kimi-K2-Thinking.md) |
| GLM-4 | ❌ | [#2255](https://github.com/vllm-project/vllm-ascend/issues/2255) |||||||||||||||||||
| GLM-4-0414 | ❌ | [#2258](https://github.com/vllm-project/vllm-ascend/issues/2258) |||||||||||||||||||
| ChatGLM | ❌ | [#554](https://github.com/vllm-project/vllm-ascend/issues/554) |||||||||||||||||||
| DeepSeek V2.5 | 🟡 | Need test |||||||||||||||||||
| Mllama | 🟡 | Need test |||||||||||||||||||
| MiniMax-Text | 🟡 | Need test |||||||||||||||||||
#### Extended Compatible Models
| Model | Support | Note | Supported Hardware |
|-------------------------------|-----------|----------------------------------------------------------------------|--------------------|
| DeepSeek Distill (Qwen/Llama) | ✅ | | A2/A3 |
| Qwen3-based | ✅ | | A2/A3 |
| Qwen2 | ✅ | | A2/A3 |
| Qwen2-based | ✅ | | A2/A3 |
| QwQ-32B | ✅ | | A2/A3 |
| Llama2/3/3.1/3.2 | ✅ | | A2/A3 |
| Internlm | ✅ | [#1962](https://github.com/vllm-project/vllm-ascend/issues/1962) | A2/A3 |
| Baichuan | ✅ | | A2/A3 |
| Baichuan2 | ✅ | | A2/A3 |
| Phi-4-mini | ✅ | | A2/A3 |
| MiniCPM | ✅ | | A2/A3 |
| MiniCPM3 | ✅ | | A2/A3 |
| Ernie4.5 | ✅ | | A2/A3 |
| Ernie4.5-Moe | ✅ | | A2/A3 |
| Gemma-2 | ✅ | | A2/A3 |
| Gemma-3 | ✅ | | A2/A3 |
| Phi-3/4 | ✅ | | A2/A3 |
| Mistral/Mistral-Instruct | ✅ | | A2/A3 |
| GLM-4 | ❌ | [#2255](https://github.com/vllm-project/vllm-ascend/issues/2255) | |
| GLM-4-0414 | ❌ | [#2258](https://github.com/vllm-project/vllm-ascend/issues/2258) | |
| ChatGLM | ❌ | [#554](https://github.com/vllm-project/vllm-ascend/issues/554) | |
| DeepSeek V2.5 | 🟡 | Need test | |
| Mllama | 🟡 | Need test | |
| MiniMax-Text | 🟡 | Need test | |
### Pooling Models
| Model | Support | Note | BF16 | Supported Hardware | W8A8 | Chunked Prefill | Automatic Prefix Cache | LoRA | Speculative Decoding | Async Scheduling | Tensor Parallel | Pipeline Parallel | Expert Parallel | Data Parallel | Prefill-decode Disaggregation | Piecewise AclGraph | Fullgraph AclGraph | max-model-len | MLP Weight Prefetch | Doc |
|-------------------------------|-----------|----------------------------------------------------------------------|------|--------------------|------|-----------------|------------------------|------|----------------------|------------------|-----------------|-------------------|-----------------|---------------|-------------------------------|--------------------|--------------------|---------------|---------------------|-----|
| Qwen3-Embedding | ✅ | || A2/A3 |||||||||||||||| [Qwen3_embedding](../../tutorials/Qwen3_embedding.md)|
| Qwen3-Reranker | ✅ | || A2/A3 |||||||||||||||| [Qwen3_reranker](../../tutorials/Qwen3_reranker.md)|
| Molmo | ✅ | [1942](https://github.com/vllm-project/vllm-ascend/issues/1942) || A2/A3 |||||||||||||||||
| XLM-RoBERTa-based | ✅ | || A2/A3 |||||||||||||||||
| Bert | ✅ | || A2/A3 |||||||||||||||||
| Model | Support | Note | Supported Hardware | Doc |
|-------------------------------|-----------|----------------------------------------------------------------------|--------------------------|------|
| Qwen3-Embedding | ✅ | | A2/A3 | [Qwen3_embedding](../../tutorials/Qwen3_embedding.md)|
| Qwen3-Reranker | ✅ | | A2/A3 | [Qwen3_reranker](../../tutorials/Qwen3_reranker.md)|
| Molmo | ✅ | [1942](https://github.com/vllm-project/vllm-ascend/issues/1942) | A2/A3 | |
| XLM-RoBERTa-based | ✅ | | A2/A3 | |
| Bert | ✅ | | A2/A3 | |
## Multimodal Language Models
### Generative Models
#### Core Supported Models
| Model | Support | Note | BF16 | Supported Hardware | W8A8 | Chunked Prefill | Automatic Prefix Cache | LoRA | Speculative Decoding | Async Scheduling | Tensor Parallel | Pipeline Parallel | Expert Parallel | Data Parallel | Prefill-decode Disaggregation | Piecewise AclGraph | Fullgraph AclGraph | max-model-len | MLP Weight Prefetch | Doc |
|--------------------------------|---------------|----------------------------------------------------------------------|------|--------------------|------|-----------------|------------------------|------|----------------------|------------------|-----------------|-------------------|-----------------|---------------|-------------------------------|--------------------|--------------------|---------------|---------------------|-----|
| Qwen2-VL | ✅ | || A2/A3 |||||||||||||||||
| Qwen2.5-VL | ✅ | | ✅ | A2/A3 | ✅ | ✅ | ✅ ||| ✅ | ✅ |||| ✅ | ✅ | ✅ | 30k || [Qwen-VL-Dense](../../tutorials/Qwen-VL-Dense.md) |
| Qwen3-VL | ✅ | ||A2/A3|||||||✅|||||✅|✅||| [Qwen-VL-Dense](../../tutorials/Qwen-VL-Dense.md) |
| Qwen3-VL-MOE | ✅ | | ✅ | A2/A3||✅|✅|||✅|✅|✅|✅|✅|✅|✅|✅|256k||[Qwen3-VL-MOE](../../tutorials/Qwen3-VL-235B-A22B-Instruct.md)|
| Qwen3-Omni-30B-A3B-Thinking | ✅ | ||A2/A3|||||||✅||✅|||||||[Qwen3-Omni-30B-A3B-Thinking](../../tutorials/Qwen3-Omni-30B-A3B-Thinking.md)|
| Qwen2.5-Omni | ✅ | || A2/A3 |||||||||||||||| [Qwen2.5-Omni](../../tutorials/Qwen2.5-Omni.md) |
| Qwen3-Omni | ✅ | || A2/A3 |||||||||||||||||
| QVQ | ✅ | || A2/A3 |||||||||||||||||
| Qwen2-Audio | ✅ | || A2/A3 |||||||||||||||||
| Aria | | || A2/A3 |||||||||||||||||
| LLaVA-Next | ✅ | || A2/A3 |||||||||||||||||
| LLaVA-Next-Video | ✅ | || A2/A3 |||||||||||||||||
| MiniCPM-V | ✅ | || A2/A3 |||||||||||||||||
| Mistral3 | ✅ | || A2/A3 |||||||||||||||||
| Phi-3-Vision/Phi-3.5-Vision | ✅ | || A2/A3 |||||||||||||||||
| Gemma3 | ✅ | || A2/A3 |||||||||||||||||
| Llama3.2 | ✅ | || A2/A3 |||||||||||||||||
| PaddleOCR-VL | ✅ | || A2/A3 |||||||||||||||||
| Llama4 | | [1972](https://github.com/vllm-project/vllm-ascend/issues/1972) |||||||||||||||||||
| Keye-VL-8B-Preview | | [1963](https://github.com/vllm-project/vllm-ascend/issues/1963) |||||||||||||||||||
| Florence-2 | ❌ | [2259](https://github.com/vllm-project/vllm-ascend/issues/2259) |||||||||||||||||||
| GLM-4V | | [2260](https://github.com/vllm-project/vllm-ascend/issues/2260) |||||||||||||||||||
| InternVL2.0/2.5/3.0<br>InternVideo2.5/Mono-InternVL | ❌ | [2064](https://github.com/vllm-project/vllm-ascend/issues/2064) |||||||||||||||||||
| Whisper | ❌ | [2262](https://github.com/vllm-project/vllm-ascend/issues/2262) |||||||||||||||||||
| Ultravox | 🟡 | Need test |||||||||||||||||||
#### Extended Compatible Models
| Model | Support | Note | Supported Hardware |
|--------------------------------|---------------|----------------------------------------------------------------------|--------------------|
| Qwen2-VL | ✅ | | A2/A3 |
| Qwen3-Omni | ✅ | | A2/A3 |
| QVQ | ✅ | | A2/A3 |
| Qwen2-Audio | ✅ | | A2/A3 |
| Aria | ✅ | | A2/A3 |
| LLaVA-Next | ✅ | | A2/A3 |
| LLaVA-Next-Video | ✅ | | A2/A3 |
| MiniCPM-V | ✅ | | A2/A3 |
| Mistral3 | | | A2/A3 |
| Phi-3-Vision/Phi-3.5-Vision | ✅ | | A2/A3 |
| Gemma3 | | | A2/A3 |
| Llama3.2 | ✅ | | A2/A3 |
| PaddleOCR-VL | ✅ | | A2/A3 |
| Llama4 | | [1972](https://github.com/vllm-project/vllm-ascend/issues/1972) | |
| Keye-VL-8B-Preview | ❌ | [1963](https://github.com/vllm-project/vllm-ascend/issues/1963) | |
| Florence-2 | ❌ | [2259](https://github.com/vllm-project/vllm-ascend/issues/2259) | |
| GLM-4V | ❌ | [2260](https://github.com/vllm-project/vllm-ascend/issues/2260) | |
| InternVL2.0/2.5/3.0<br>InternVideo2.5/Mono-InternVL | ❌ | [2064](https://github.com/vllm-project/vllm-ascend/issues/2064) | |
| Whisper | ❌ | [2262](https://github.com/vllm-project/vllm-ascend/issues/2262) | |
| Ultravox | 🟡 | Need test | |