Update the transformers version in CI (#1690)

This commit is contained in:
Lianmin Zheng
2024-10-16 19:03:55 -07:00
committed by GitHub
parent 2782132be8
commit 02f7f3e488
2 changed files with 11 additions and 9 deletions

View File

@@ -29,7 +29,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[dev]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Run test
@@ -49,7 +49,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[dev]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Run test
@@ -69,7 +69,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[dev]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Run test
@@ -89,7 +89,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[dev]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Run test
@@ -109,7 +109,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[all]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Benchmark Single Latency
@@ -147,7 +147,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[all]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Benchmark Offline Throughput (w/o RadixAttention)
@@ -179,7 +179,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[all]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
- name: Benchmark Offline Throughput (TP=2)
@@ -211,7 +211,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[all]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
git clone https://github.com/merrymercy/human-eval.git
@@ -235,7 +235,7 @@ jobs:
run: |
pip install --upgrade pip
pip install -e "python[all]"
pip install transformers==4.44
pip install transformers==4.45.2
pip install flashinfer -i https://flashinfer.ai/whl/cu121/torch2.4/ --force-reinstall
git clone https://github.com/merrymercy/human-eval.git