[Test] Add flashmla attention backend test (#5587)
This commit is contained in:
@@ -51,6 +51,7 @@ suites = {
|
||||
TestFile("test_mla_int8_deepseek_v3.py", 389),
|
||||
TestFile("test_mla_flashinfer.py", 395),
|
||||
TestFile("test_mla_fp8.py", 153),
|
||||
TestFile("test_flash_mla_attention_backend.py", 300),
|
||||
TestFile("test_no_chunked_prefill.py", 108),
|
||||
TestFile("test_no_overlap_scheduler.py", 216),
|
||||
TestFile("test_openai_server.py", 149),
|
||||
|
||||
Reference in New Issue
Block a user