[Revision] Replace enable_flashinfer_mla argument with attention_backend (#5052)
This commit is contained in:
@@ -26,7 +26,8 @@ class TestFlashinferMLA(CustomTestCase):
|
||||
"--enable-torch-compile",
|
||||
"--cuda-graph-max-bs",
|
||||
"2",
|
||||
"--enable-flashinfer-mla",
|
||||
"--attention-backend",
|
||||
"flashinfer",
|
||||
]
|
||||
)
|
||||
cls.process = popen_launch_server(
|
||||
@@ -69,8 +70,8 @@ class TestFlashinferMLANoRagged(CustomTestCase):
|
||||
"--disable-cuda-graph",
|
||||
"--cuda-graph-max-bs",
|
||||
"4",
|
||||
"--enable-flashinfer-mla",
|
||||
"--flashinfer-mla-disable-ragged",
|
||||
"--attention-backend",
|
||||
"flashinfer",
|
||||
]
|
||||
)
|
||||
cls.process = popen_launch_server(
|
||||
@@ -125,7 +126,8 @@ class TestFlashinferMLAMTP(CustomTestCase):
|
||||
"1",
|
||||
"--speculative-num-draft-tokens",
|
||||
"4",
|
||||
"--enable-flashinfer-mla",
|
||||
"--attention-backend",
|
||||
"flashinfer",
|
||||
]
|
||||
)
|
||||
cls.process = popen_launch_server(
|
||||
|
||||
Reference in New Issue
Block a user