diff --git a/lib/Conversion/TritonGPUToLLVM/DotHelpers.h b/lib/Conversion/TritonGPUToLLVM/DotHelpers.h index 728857d39..d354f0f4d 100644 --- a/lib/Conversion/TritonGPUToLLVM/DotHelpers.h +++ b/lib/Conversion/TritonGPUToLLVM/DotHelpers.h @@ -1138,7 +1138,7 @@ private: std::function getLoadMatrixFn(Value tensor, const SharedMemoryObject &smemObj, MmaEncodingAttr mmaLayout, int wpt, uint32_t kOrder, - ArrayRef instrShape, ArrayRef matShape, + SmallVector instrShape, SmallVector matShape, Value warpId, ValueTable &vals, bool isA) const { auto tensorTy = tensor.getType().cast(); // We assumes that the input operand of Dot should be from shared layout. diff --git a/lib/Conversion/TritonGPUToLLVM/TritonGPUToLLVM.cpp b/lib/Conversion/TritonGPUToLLVM/TritonGPUToLLVM.cpp index 0a290fff1..8253d1786 100644 --- a/lib/Conversion/TritonGPUToLLVM/TritonGPUToLLVM.cpp +++ b/lib/Conversion/TritonGPUToLLVM/TritonGPUToLLVM.cpp @@ -1875,9 +1875,9 @@ struct PrintfOpConversion Value globalPtr = rewriter.create(UnknownLoc::get(context), global); - Value stringStart = - rewriter.create(UnknownLoc::get(context), int8Ptr, - globalPtr, mlir::ValueRange({zero, zero})); + Value stringStart = rewriter.create( + UnknownLoc::get(context), int8Ptr, globalPtr, + SmallVector({zero, zero})); Value bufferPtr = rewriter.create(UnknownLoc::get(context), int8Ptr); @@ -1912,7 +1912,7 @@ struct PrintfOpConversion int8Ptr, allocated); } - ValueRange operands{stringStart, bufferPtr}; + SmallVector operands{stringStart, bufferPtr}; rewriter.create(UnknownLoc::get(context), funcOp, operands); } }; diff --git a/python/setup.py b/python/setup.py index 73b03d55f..ef0353059 100644 --- a/python/setup.py +++ b/python/setup.py @@ -25,8 +25,8 @@ def get_build_type(): elif check_env_flag("REL_WITH_DEB_INFO"): return "RelWithDebInfo" else: - return "Debug" - # TODO(Keren): Restore this before we merge into master + return "RelWithDebInfo" + # TODO: change to release when stable enough #return "Release"