Support sm90 Int8 gemm (#3035)
This commit is contained in:
@@ -25,7 +25,7 @@ class TestInt8Gemm(unittest.TestCase):
|
||||
scale_a = torch.randn((M,), device="cuda", dtype=torch.float32)
|
||||
scale_b = torch.randn((N,), device="cuda", dtype=torch.float32)
|
||||
if with_bias:
|
||||
bias = torch.ones((N,), device="cuda", dtype=out_dtype) * 10
|
||||
bias = torch.randn((N,), device="cuda", dtype=out_dtype) * 10
|
||||
else:
|
||||
bias = None
|
||||
|
||||
|
||||
Reference in New Issue
Block a user