chore: update vllm to 0.5.4 (#966)
This commit is contained in:
3
.github/workflows/e2e-test.yml
vendored
3
.github/workflows/e2e-test.yml
vendored
@@ -34,8 +34,7 @@ jobs:
|
||||
pip cache purge
|
||||
pip install --upgrade pip
|
||||
pip install -e "python[all]"
|
||||
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3/ --force-reinstall
|
||||
pip install --upgrade transformers
|
||||
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
|
||||
|
||||
- name: Benchmark Serving Throughput
|
||||
run: |
|
||||
|
||||
3
.github/workflows/unit-test.yml
vendored
3
.github/workflows/unit-test.yml
vendored
@@ -34,8 +34,7 @@ jobs:
|
||||
pip cache purge
|
||||
pip install --upgrade pip
|
||||
pip install -e "python[all]"
|
||||
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.3/ --force-reinstall
|
||||
pip install --upgrade transformers
|
||||
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
|
||||
pip install accelerate
|
||||
|
||||
- name: Test Frontend Language
|
||||
|
||||
Reference in New Issue
Block a user