triton-lang / triton

Development repository for the Triton language and compiler
https://triton-lang.org/
MIT License
12.8k stars 1.54k forks source link

ROCm 5.5 "cumputeCapability" issue #1634

Open cornpo opened 1 year ago

cornpo commented 1 year ago

Hello,

With ROCm 5.5, 2.1.0.dev20230502+rocm5.4.2, Triton 2.0.0.post1, pytorch-triton-rocm 2.0.2

Running oobabooga text webui loading a quantized model.

python: /project/lib/Dialect/TritonGPU/Transforms/Combine.cpp:870: int {anonymous}::{anonymous}::computeCapabilityToMMAVersion(int): Assertionfalse && "computeCapability > 90 not supported"' failed.`

It works with cuda. It will load the model in Triton. But gives that error at infer time.

psinger commented 1 year ago

@cornpo running into the same, did you find a solution?

cornpo commented 1 year ago

Not for the Triton stuff. Cuda and old-cuda worked, as I remember. Haven't tried anything new since beginning of June.

Good luck.

phaylali commented 1 year ago

anyone found a solution for this?

binarman commented 1 year ago

Hi everyone!

Triton 2.0 is relatively old version of triton, is it mandatory to use it for you?