Closed nonetrix closed 4 months ago
I have also reproduced this with SD 1.5, and SDXL model which is not based on PonyXL which is slightly different apparently. It does allocate VRAM, but then just crashes
Try this before launching: export HSA_OVERRIDE_GFX_VERSION=10.3.0
It's a common workaround for the python version of SD, maybe it works in the same way
Hm never had to do that tbh maybe it's automatic though most of the time, I'll try that
Hm never had to do that tbh maybe it's automatic though most of the time, I'll try that
In automatic1111's webui there are some lines in the start script that do that for you, as well in some apps like koboldcpp. .....but in theory it shouldn't be needed for your gpu.
Also, if you followed the guide in homepage, probably you have to recompile using a different DAMDGPU_TARGETS parameter. gfx1100 is for the 7000 series, while the codename for your gpu is gfx1030 (you can get those here, it's a bit outdated but still useful https://llvm.org/docs/AMDGPUUsage.html).
Then, try to do
cmake .. -G "Ninja" -DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++ -DSD_HIPBLAS=ON -DCMAKE_BUILD_TYPE=Release -DAMDGPU_TARGETS=gfx1030
cmake --build . --config Release
...Or maybe it's a completely different issue, but trying does not harm
Hm never had to do that tbh maybe it's automatic though most of the time, I'll try that
In automatic1111's webui there are some lines in the start script that do that for you, as well in some apps like koboldcpp. .....but in theory it shouldn't be needed for your gpu.
Also, if you followed the guide in homepage, probably you have to recompile using a different DAMDGPU_TARGETS parameter. gfx1100 is for the 7000 series, while the codename for your gpu is gfx1030 (you can get those here, it's a bit outdated but still useful https://llvm.org/docs/AMDGPUUsage.html).
Then, try to do
cmake .. -G "Ninja" -DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++ -DSD_HIPBLAS=ON -DCMAKE_BUILD_TYPE=Release -DAMDGPU_TARGETS=gfx1030 cmake --build . --config Release
...Or maybe it's a completely different issue, but trying does not harm
That worked for me and is a lot faster, thanks. Maybe this should be added to the read me though? Also as a side effect gets around my issue with the PyTorch backend where my GPU will reset annoyingly due to a firmware blob bug that has been fixed but not released. That is unrelated mostly, but it only happens with ROCm or Vulkan compute which is rather annoying, and I've only triggered it on Vulkan llama cpp and any Stable Diffusion UI using PyTorch, but then it's fine with PyTorch LLMs it's a annoying mess just wanted to rant lol. I think I might make a terminal client for this if it isn't too hard, but I think I'll use iTerm image support which is also in Wezterm maybe add Kitty support too or other terminals not sure. Is there some kind of http API I could plugin to btw?
Also my GPU is so quiet when using this compared to PyTorch the fans barley spin up meanwhile PyTorch they max out and it's much slower
Also my GPU is so quiet when using this compared to PyTorch the fans barley spin up meanwhile PyTorch they max out and it's much slower
cpp version is much slower or the PyTorch version?
PyTorch sorry for my garbage wording and grammar
I am trying to run on my RX 6800 on Arch Linux distrobox running inside of Arch Linux but am getting this error when I try to generate anything
It looks like I am missing some dependency but I installed pretty much all the ROCm packages and Hipblas etc.