chore: upgrade flashinfer v0.2.6.post1 jit (#6958)

Co-authored-by: alcanderian <alcanderian@gmail.com>
Co-authored-by: Qiaolin Yu <qy254@cornell.edu>
Co-authored-by: Baizhou Zhang <sobereddiezhang@gmail.com>
Co-authored-by: Mick <mickjagger19@icloud.com>
Co-authored-by: ispobock <ispobaoke@gmail.com>
This commit is contained in:
Yineng Zhang
2025-06-09 09:22:39 -07:00
committed by GitHub
parent 98c00a2df1
commit 56ccd3c22c
14 changed files with 189 additions and 27 deletions

View File

@@ -10,8 +10,8 @@ bash "${SCRIPT_DIR}/killall_sglang.sh"
pip install --upgrade pip
# Clean up existing installations
pip uninstall -y flashinfer flashinfer_python sgl-kernel sglang vllm
pip cache purge
pip uninstall -y flashinfer flashinfer_python sgl-kernel sglang vllm || true
pip cache purge || true
rm -rf /root/.cache/flashinfer
rm -rf /usr/local/lib/python3.10/dist-packages/flashinfer*
rm -rf /usr/local/lib/python3.10/dist-packages/sgl_kernel*
@@ -19,6 +19,9 @@ rm -rf /usr/local/lib/python3.10/dist-packages/sgl_kernel*
# Install the main package
pip install -e "python[dev]"
# Show current packages
pip list
# Install additional dependencies
pip install mooncake-transfer-engine==0.3.2.post1 nvidia-cuda-nvrtc-cu12
@@ -27,7 +30,13 @@ git clone --branch v0.3.3 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eva
pip install -e lmms-eval/
# Install FlashMLA for attention backend tests
pip install git+https://github.com/deepseek-ai/FlashMLA.git
# pip install git+https://github.com/deepseek-ai/FlashMLA.git
# Install hf_xet
pip install huggingface_hub[hf_xet]
# Install xformers
pip install -U xformers --index-url https://download.pytorch.org/whl/cu126 --no-deps --force-reinstall
# Show current packages
pip list