chore: update vllm to 0.5.4 (#966)

This commit is contained in:
Yineng Zhang
2024-08-07 19:15:41 +08:00
committed by GitHub
parent a01ddd9605
commit c31f084c71
14 changed files with 15 additions and 18 deletions

View File

@@ -34,8 +34,7 @@ jobs:
pip cache purge
pip install --upgrade pip
pip install -e "python[all]"
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3/ --force-reinstall
pip install --upgrade transformers
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Benchmark Serving Throughput
run: |

View File

@@ -34,8 +34,7 @@ jobs:
pip cache purge
pip install --upgrade pip
pip install -e "python[all]"
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3/ --force-reinstall
pip install --upgrade transformers
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
pip install accelerate
- name: Test Frontend Language