chore: update vllm to 0.5.4 (#966)

This commit is contained in:
Yineng Zhang
2024-08-07 19:15:41 +08:00
committed by GitHub
parent a01ddd9605
commit c31f084c71
14 changed files with 15 additions and 18 deletions

View File

@@ -49,7 +49,7 @@ pip install --upgrade pip
pip install "sglang[all]"
# Install FlashInfer CUDA kernels
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3/
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/
```
### Method 2: From source
@@ -62,7 +62,7 @@ pip install --upgrade pip
pip install -e "python[all]"
# Install FlashInfer CUDA kernels
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3/
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/
```
### Method 3: Using docker