[fix] fix misusing of is_cuda (#7790)

This commit is contained in:
JieXin Liang
2025-07-05 19:02:14 +08:00
committed by GitHub
parent 6c903611ca
commit c04a8a820b

View File

@@ -913,7 +913,7 @@ class DeepseekV2AttentionMLA(nn.Module):
and self.fused_qkv_a_proj_with_mqa.weight.dtype == torch.bfloat16
and self.fused_qkv_a_proj_with_mqa.weight.shape[0] == 2112
and self.fused_qkv_a_proj_with_mqa.weight.shape[1] == 7168
and is_cuda
and _is_cuda
and _device_sm >= 90
)