[PYTHON] Fixed issue with IS_TK_DIV_K
This commit is contained in:
@@ -57,7 +57,7 @@ class _matmul(torch.autograd.Function):
|
|||||||
'TN' : _matmul.TN,
|
'TN' : _matmul.TN,
|
||||||
'TK' : _matmul.TK,
|
'TK' : _matmul.TK,
|
||||||
'TZ' : _matmul.TZ,
|
'TZ' : _matmul.TZ,
|
||||||
'IS_TK_DIV_K' : is_tk_div_k
|
'IS_TK_DIV_K' : int(is_tk_div_k)
|
||||||
}
|
}
|
||||||
_matmul._kernels[key] = triton.kernel(_matmul.src, device, num_warps=_matmul.num_warps, defines=defines)
|
_matmul._kernels[key] = triton.kernel(_matmul.src, device, num_warps=_matmul.num_warps, defines=defines)
|
||||||
kernel = _matmul._kernels[key]
|
kernel = _matmul._kernels[key]
|
||||||
|
Reference in New Issue
Block a user