chore: upgrade flashinfer v0.2.6.post1 jit (#6958)
Co-authored-by: alcanderian <alcanderian@gmail.com> Co-authored-by: Qiaolin Yu <qy254@cornell.edu> Co-authored-by: Baizhou Zhang <sobereddiezhang@gmail.com> Co-authored-by: Mick <mickjagger19@icloud.com> Co-authored-by: ispobock <ispobaoke@gmail.com>
This commit is contained in:
@@ -10,8 +10,8 @@ bash "${SCRIPT_DIR}/killall_sglang.sh"
|
||||
pip install --upgrade pip
|
||||
|
||||
# Clean up existing installations
|
||||
pip uninstall -y flashinfer flashinfer_python sgl-kernel sglang vllm
|
||||
pip cache purge
|
||||
pip uninstall -y flashinfer flashinfer_python sgl-kernel sglang vllm || true
|
||||
pip cache purge || true
|
||||
rm -rf /root/.cache/flashinfer
|
||||
rm -rf /usr/local/lib/python3.10/dist-packages/flashinfer*
|
||||
rm -rf /usr/local/lib/python3.10/dist-packages/sgl_kernel*
|
||||
@@ -19,6 +19,9 @@ rm -rf /usr/local/lib/python3.10/dist-packages/sgl_kernel*
|
||||
# Install the main package
|
||||
pip install -e "python[dev]"
|
||||
|
||||
# Show current packages
|
||||
pip list
|
||||
|
||||
# Install additional dependencies
|
||||
pip install mooncake-transfer-engine==0.3.2.post1 nvidia-cuda-nvrtc-cu12
|
||||
|
||||
@@ -27,7 +30,13 @@ git clone --branch v0.3.3 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eva
|
||||
pip install -e lmms-eval/
|
||||
|
||||
# Install FlashMLA for attention backend tests
|
||||
pip install git+https://github.com/deepseek-ai/FlashMLA.git
|
||||
# pip install git+https://github.com/deepseek-ai/FlashMLA.git
|
||||
|
||||
# Install hf_xet
|
||||
pip install huggingface_hub[hf_xet]
|
||||
|
||||
# Install xformers
|
||||
pip install -U xformers --index-url https://download.pytorch.org/whl/cu126 --no-deps --force-reinstall
|
||||
|
||||
# Show current packages
|
||||
pip list
|
||||
|
||||
Reference in New Issue
Block a user