Closed binarman closed 5 months ago
Converted to draft, because first want to merge mfma4x64 support, and tests generating scripts (for accelerate matmul lit tests)
@alefimov-amd @scxiao I'd suggest we close this one. The type promotion is upstream'ed and we don't need it in terms of perf.
@zhanglx13 I want to update this PR, when upstream is stable(at least one more PR: https://github.com/openai/triton/pull/3025), to keep upstream and triton-mlir compatible.
Eventually, we will need to transfer all the stuff we have in triton-mlir, if upstream and triton-mlir are compatible, we can simply copy code from triton-mlir to upstream.
@alefimov-amd Ok, I'm also fine with that
This PR did 3 things:
1, Move C/D operand casts from python to C++ code
Now this PR is doing only 1 item. 2 item is moved to #496 3 item is temporary removed, since mixed precision is not supported in tt.dot operation(MLIR verification fails), I want to discuss this with upstream first
This PR moves casts of operands from python code and ttg to llvm phase to AccelerateAMDMatmul pass