From f64b779b0de8d3b7652f33c1848a21e6067ed190 Mon Sep 17 00:00:00 2001 From: Philippe Tillet Date: Wed, 24 Feb 2021 14:43:08 -0500 Subject: [PATCH] [PYTHON] Bugfix on FP32 blocksparse matmul --- python/triton/ops/blocksparse/matmul.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/triton/ops/blocksparse/matmul.py b/python/triton/ops/blocksparse/matmul.py index e874eee89..3ad63dcd3 100644 --- a/python/triton/ops/blocksparse/matmul.py +++ b/python/triton/ops/blocksparse/matmul.py @@ -362,7 +362,7 @@ class matmul: return self.lut_cache[key] # C look-up table layout, block = self.layout, self.block - step = 8 if dtype == torch.float32 else 16 + step = 16 if self.mode == 'sdd': c_lut, c_num_locks, c_width, c_packs = _matmul.make_sdd_lut(layout, block, dtype, device) elif self.mode == 'dsd':