[Test] Add flashmla attention backend test (#5587)

This commit is contained in:
Huapeng Zhou
2025-05-06 01:32:02 +08:00
committed by GitHub
parent 56f6589ecb
commit b8559764f6
3 changed files with 68 additions and 0 deletions

View File

@@ -51,6 +51,7 @@ suites = {
TestFile("test_mla_int8_deepseek_v3.py", 389),
TestFile("test_mla_flashinfer.py", 395),
TestFile("test_mla_fp8.py", 153),
TestFile("test_flash_mla_attention_backend.py", 300),
TestFile("test_no_chunked_prefill.py", 108),
TestFile("test_no_overlap_scheduler.py", 216),
TestFile("test_openai_server.py", 149),