Replace enable_flashinfer_mla argument with attention_backend (#5005)

This commit is contained in:
Baizhou Zhang
2025-04-03 02:53:58 -07:00
committed by GitHub
parent 772d2a191d
commit e8999b13b7
8 changed files with 21 additions and 20 deletions

View File

@@ -26,7 +26,8 @@ class TestFlashinferMLA(CustomTestCase):
"--enable-torch-compile",
"--cuda-graph-max-bs",
"2",
"--enable-flashinfer-mla",
"--attention-backend",
"flashinfer",
]
)
cls.process = popen_launch_server(
@@ -69,8 +70,8 @@ class TestFlashinferMLANoRagged(CustomTestCase):
"--disable-cuda-graph",
"--cuda-graph-max-bs",
"4",
"--enable-flashinfer-mla",
"--flashinfer-mla-disable-ragged",
"--attention-backend",
"flashinfer",
]
)
cls.process = popen_launch_server(
@@ -125,7 +126,8 @@ class TestFlashinferMLAMTP(CustomTestCase):
"1",
"--speculative-num-draft-tokens",
"4",
"--enable-flashinfer-mla",
"--attention-backend",
"flashinfer",
]
)
cls.process = popen_launch_server(