Disable kernel cutlass_mla_decode on SM103 (#10058)
Signed-off-by: Hao Lu <14827759+hlu1@users.noreply.github.com>
This commit is contained in:
@@ -4,9 +4,10 @@ import torch.nn.functional as F
|
||||
from sgl_kernel import cutlass_mla_decode, cutlass_mla_get_workspace_size
|
||||
from torch import Tensor
|
||||
|
||||
if torch.cuda.get_device_capability() < (10, 0):
|
||||
# Disable tests on SM103 until the accuracy issues are fixed.
|
||||
if torch.cuda.get_device_capability() != (10, 0):
|
||||
pytest.skip(
|
||||
reason="Cutlass MLA Requires compute capability of 10 or above.",
|
||||
reason="Cutlass MLA Requires compute capability of 10.",
|
||||
allow_module_level=True,
|
||||
)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user