Merge pull request #19 from ROCmSoftwarePlatform/unskip_test_reduce

reduce the skips for test_reduce functions
This commit is contained in:
rsanthanam-amd
2022-11-01 11:05:18 -05:00
committed by GitHub

View File

@@ -892,7 +892,8 @@ def test_f16_to_f8_rounding():
def test_reduce1d(op, dtype_str, shape, device='cuda'):
check_type_supported(dtype_str) # bfloat16 on cc < 80 will not be tested
if torch.version.hip is not None:
pytest.skip(f"test_reduce1d currently has segfaults on ROCM")
if dtype_str in ["int8", "int16", "uint8", "uint16"]:
pytest.skip(f"test_reduce1d[{dtype_str}] skipped on ROCM")
# triton kernel
@triton.jit
@@ -953,7 +954,8 @@ reduce_configs2 = [
def test_reduce2d(op, dtype_str, shape, axis, device='cuda'):
check_type_supported(dtype_str) # bfloat16 on cc < 80 will not be tested
if torch.version.hip is not None:
pytest.skip(f"test_reduce2d currently has segfaults on ROCM")
if dtype_str in ["int8", "int16", "uint8", "uint16"]:
pytest.skip(f"test_reduce2d[{dtype_str}] skipped on ROCM")
# triton kernel
@triton.jit
def kernel(X, Z, BLOCK_M: tl.constexpr, BLOCK_N: tl.constexpr, AXIS: tl.constexpr):