2024-12-31 11:04:01 +08:00
|
|
|
#!/bin/bash
|
2024-12-01 18:55:26 +08:00
|
|
|
# Install the dependency in CI.
|
2025-03-28 10:34:10 -07:00
|
|
|
set -euxo pipefail
|
2024-10-30 02:49:08 -07:00
|
|
|
|
2025-03-28 10:34:10 -07:00
|
|
|
# Use repo from environment variables, passed from GitHub Actions
|
2025-02-14 08:50:14 +08:00
|
|
|
FLASHINFER_REPO="${FLASHINFER_REPO:-https://flashinfer.ai/whl/cu124/torch2.5/flashinfer-python}"
|
2024-12-01 18:55:26 +08:00
|
|
|
|
|
|
|
|
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
|
|
|
|
|
bash "${SCRIPT_DIR}/killall_sglang.sh"
|
2024-11-30 00:24:30 -08:00
|
|
|
|
2024-10-26 04:32:36 -07:00
|
|
|
pip install --upgrade pip
|
2025-02-04 21:41:40 +08:00
|
|
|
pip uninstall flashinfer -y
|
2025-02-14 08:50:14 +08:00
|
|
|
pip install -e "python[all]" --find-links https://flashinfer.ai/whl/cu124/torch2.5/flashinfer-python
|
2024-12-01 01:47:30 -08:00
|
|
|
|
2025-02-14 08:50:14 +08:00
|
|
|
rm -rf /root/.cache/flashinfer
|
2025-01-14 03:38:51 +08:00
|
|
|
# Force reinstall flashinfer and torch_memory_saver
|
2025-03-11 15:37:17 -07:00
|
|
|
pip install flashinfer_python==0.2.3 --find-links ${FLASHINFER_REPO} --force-reinstall --no-deps
|
2025-03-28 10:34:10 -07:00
|
|
|
pip install sgl-kernel==0.0.5.post3 --force-reinstall
|
2025-02-14 08:50:14 +08:00
|
|
|
|
2025-03-28 10:34:10 -07:00
|
|
|
pip install torch_memory_saver
|
|
|
|
|
pip install transformers==4.50.0 sentence_transformers accelerate==1.4.0 peft pandas datasets timm
|
2024-12-01 01:47:30 -08:00
|
|
|
|
|
|
|
|
# For compling xgrammar kernels
|
2024-11-25 04:58:16 -08:00
|
|
|
pip install cuda-python nvidia-cuda-nvrtc-cu12
|
2025-03-17 14:07:59 +08:00
|
|
|
|
2025-03-27 04:21:25 +08:00
|
|
|
pip uninstall vllm -y || true
|