Closed markg85 closed 1 year ago
I'm less familiar with AMD and rocm side of things. Do you see have any way of checking gpu utilization while running the h2ogpt model? I'd like to double check if it is actually running on gpu or if it is just pushing it to cpu.
Seems to me like if it is actually working on amd gpu then it is snagging on the bitsandbytes piece that relies on cuda in a way that isn't compatible with cuda(hip)/rocm
Yeah, i resolved this. This bitsandbytes fork hacks in rocm support and makes it work.
So closing my own issue :)
Hi,
No clue if it's intentional or accidental, but i'm having issues running any model besides h2ogpt ones.
Just to be sure with regard to hardware and versions. I'm running pytorch woth rocm (AMD GPU). This does work just fine on h2ogpt models! Output of some pytorch specifics:
I'm guessing - based on the error - that it wants to look for cuda which just isn't there in a rocm environment.