[Revision] Replace enable_flashinfer_mla argument with attention_backend (#5052)

This commit is contained in:
Baizhou Zhang
2025-04-05 01:23:02 -07:00
committed by GitHub
parent ca8d02abd5
commit efbae697b3
9 changed files with 92 additions and 82 deletions

View File

@@ -26,7 +26,8 @@ class TestFlashinferMLA(CustomTestCase):
"--enable-torch-compile",
"--cuda-graph-max-bs",
"2",
"--enable-flashinfer-mla",
"--attention-backend",
"flashinfer",
]
)
cls.process = popen_launch_server(
@@ -69,8 +70,8 @@ class TestFlashinferMLANoRagged(CustomTestCase):
"--disable-cuda-graph",
"--cuda-graph-max-bs",
"4",
"--enable-flashinfer-mla",
"--flashinfer-mla-disable-ragged",
"--attention-backend",
"flashinfer",
]
)
cls.process = popen_launch_server(
@@ -125,7 +126,8 @@ class TestFlashinferMLAMTP(CustomTestCase):
"1",
"--speculative-num-draft-tokens",
"4",
"--enable-flashinfer-mla",
"--attention-backend",
"flashinfer",
]
)
cls.process = popen_launch_server(