[Test] Add flashmla attention backend test (#5587)
This commit is contained in:
@@ -31,3 +31,6 @@ pip install cuda-python nvidia-cuda-nvrtc-cu12
|
||||
# For lmms_evals evaluating MMMU
|
||||
git clone --branch v0.3.3 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
|
||||
pip install -e lmms-eval/
|
||||
|
||||
# Install FlashMLA for attention backend tests
|
||||
pip install git+https://github.com/deepseek-ai/FlashMLA.git
|
||||
|
||||
Reference in New Issue
Block a user