[MISC] fix format check error (#654)

This pr makes format.sh works as expect.

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
This commit is contained in:
wangxiyuan
2025-04-29 11:14:19 +08:00
committed by GitHub
parent 1fce70a2fb
commit 0dae55a9a3
17 changed files with 45 additions and 44 deletions

View File

@@ -89,7 +89,7 @@ Currently, w8a8 quantization is already supported by vllm-ascend originally on v
Currently, w8a8 DeepSeek is working in process: [support AscendW8A8 quantization](https://github.com/vllm-project/vllm-ascend/pull/511)
Please run DeepSeek with BF16 now, follwing the [Multi-Node DeepSeek inferencing tutorail](https://vllm-ascend.readthedocs.io/en/main/tutorials/multi_node.html)
Please run DeepSeek with BF16 now, following the [Multi-Node DeepSeek inferencing tutorail](https://vllm-ascend.readthedocs.io/en/main/tutorials/multi_node.html)
### 12. There is not output in log when loading models using vllm-ascend, How to solve it?

View File

@@ -127,7 +127,7 @@ apt update -y
apt install -y gcc g++ cmake libnuma-dev wget git
```
**[Optinal]** Config the extra-index of `pip` if you are working on a **x86** machine, so that the torch with cpu could be found:
**[Optional]** Config the extra-index of `pip` if you are working on a **x86** machine, so that the torch with cpu could be found:
```bash
pip config set global.extra-index-url https://download.pytorch.org/whl/cpu/