[PYTHON] Fixed issue with IS_TK_DIV_K
This commit is contained in:
@@ -57,7 +57,7 @@ class _matmul(torch.autograd.Function):
|
||||
'TN' : _matmul.TN,
|
||||
'TK' : _matmul.TK,
|
||||
'TZ' : _matmul.TZ,
|
||||
'IS_TK_DIV_K' : is_tk_div_k
|
||||
'IS_TK_DIV_K' : int(is_tk_div_k)
|
||||
}
|
||||
_matmul._kernels[key] = triton.kernel(_matmul.src, device, num_warps=_matmul.num_warps, defines=defines)
|
||||
kernel = _matmul._kernels[key]
|
||||
|
Reference in New Issue
Block a user