Allow use of TRTLLM_MHA backend for hybrid attention on Blackwell (#11138)
This commit is contained in:
@@ -178,7 +178,8 @@ def attn_backend_wrapper(runner, full_attn_backend):
|
||||
if is_blackwell():
|
||||
assert (
|
||||
runner.server_args.attention_backend == "triton"
|
||||
), "triton backend is the only supported backend on Blackwell GPUs for hybrid GDN models, use --attention-backend triton to specify the backend."
|
||||
or runner.server_args.attention_backend == "trtllm_mha"
|
||||
), "triton or trtllm_mha backend are the only supported backends on Blackwell GPUs for hybrid GDN models, use --attention-backend triton or --attention-backend trtllm_mha to specify the backend."
|
||||
if is_npu():
|
||||
assert (
|
||||
runner.server_args.attention_backend == "ascend"
|
||||
|
||||
Reference in New Issue
Block a user