Open gc-octopis opened 11 months ago
System: Pop!_OS 22.04 LTS x86_64 CPU: AMD Ryzen 9 5900X GPU: RX 590 8GB Environment: virtualenv Python: 3.10.12
I also got the same error. It works with --cpu
flag as a workaround but it's really slow (of course).
System: Linux Mint 21.2 Victoria base: Ubuntu 22.04 jammy CPU: AMD Ryzen 7 2700 GPU: AMD RX 590 8gb Python: 3.10.12 Environment: venv
Same error here.
Same error.
System: Ubuntu 22.04.3 LTS in WSL2 CPU: AMD Ryzen 5 2600 GPU: AMD RX 6700XT 12GB Python: 3.10.12 Environment: Venv
Same problem, I hope some solution can arrive
System: Linux Mint 21.2 Cinnamon CPU: AMD Ryzen 5 5500 GPU: AMD RX 6600M 8gb, (gpu from a notebook in a desktop pc, it's aliexpress magic). Python: 3.10.12 Environment: Venv
Ran into the same issue, fixed it by doing the following.
1 - https://docs.amd.com/en/latest/deploy/linux/quick_start.html and then setting the correct env variables for your GPU. You can check these using rocminfo. 2 - https://www.reddit.com/r/AMDHelp/comments/11l8wox/rocm_under_linux_running_on_cpu_by_default/
Facing the same issue System: Pop!_OS 22.04 CPU: Ryzen 5 7600 GPU: RX6800XT Python: Python 3.10.12 Environment: venv
Ran into the same issue, fixed it by doing the following.
1 - https://docs.amd.com/en/latest/deploy/linux/quick_start.html and then setting the correct env variables for your GPU. You can check these using rocminfo. 2 - https://www.reddit.com/r/AMDHelp/comments/11l8wox/rocm_under_linux_running_on_cpu_by_default/
Unfortunately, amdgpu-dkms doesn't seem to support kernel version 6.5 (https://github.com/RadeonOpenCompute/ROCm/issues/2458) so I was not able to install it as I'm on 6.5.6. A workaround was provided but I'll have try it out.
Facing the same issue System: Pop!_OS 22.04 CPU: Ryzen 5 7600 GPU: RX6800XT Python: Python 3.10.12 Environment: venv
Ran into the same issue, fixed it by doing the following.
1 - https://docs.amd.com/en/latest/deploy/linux/quick_start.html and then setting the correct env variables for your GPU. You can check these using rocminfo. 2 - https://www.reddit.com/r/AMDHelp/comments/11l8wox/rocm_under_linux_running_on_cpu_by_default/
Unfortunately, amdgpu-dkms doesn't seem to support kernel version 6.5 (https://github.com/RadeonOpenCompute/ROCm/issues/2458) so I was not able to install it as I'm on 6.5.6. A workaround was provided but I'll have try it out.
My apologies. I should have specified that I fixed this on Ubuntu 22.04.
Ran into the same issue, fixed it by doing the following.
1 - https://docs.amd.com/en/latest/deploy/linux/quick_start.html and then setting the correct env variables for your GPU. You can check these using rocminfo. 2 - https://www.reddit.com/r/AMDHelp/comments/11l8wox/rocm_under_linux_running_on_cpu_by_default/
I tried doing that, the 1 installed alright, but Idk how to setup the env variables, neither my rocminfo worked, neither I don't know anything anymore 🙃
What's the output when you run rocminfo.
What's the output when you run rocminfo.
it says "command 'rocminfo' not found, but it can be installed with: sudo apt install rocminfo." Then I install it but it's always the same message
In that case I assume you tried to install rocminfo with 'sudo apt install rocminfo'. This pulls it from the system repo. You need to prioritize the rocm repo. Since you're on Mint I would assume it would also work since it's basically Ubuntu / Debian.
sudo tee /etc/apt/sources.list.d/amdgpu.list <<'EOF'
deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/amdgpu/5.7.1/ubuntu jammy main
EOF
# ROCm repository for jammy
sudo tee /etc/apt/sources.list.d/rocm.list <<'EOF'
deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/debian jammy main
EOF
# Prefer packages from the rocm repository over system packages
echo -e 'Package: *\nPin: release o=repo.radeon.com\nPin-Priority: 600' | sudo tee /etc/apt/preferences.d/rocm-pin-600```
then install it with
sudo apt update
sudo apt install amdgpu-dkms
sudo apt install rocm-hip-libraries
In that case I assume you tried to install rocminfo with 'sudo apt install rocminfo'. This pulls it from the system repo. You need to prioritize the rocm repo. Since you're on Mint I would assume it would also work since it's basically Ubuntu / Debian.
sudo tee /etc/apt/sources.list.d/amdgpu.list <<'EOF' deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/amdgpu/5.7.1/ubuntu jammy main EOF # ROCm repository for jammy sudo tee /etc/apt/sources.list.d/rocm.list <<'EOF' deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/debian jammy main EOF # Prefer packages from the rocm repository over system packages echo -e 'Package: *\nPin: release o=repo.radeon.com\nPin-Priority: 600' | sudo tee /etc/apt/preferences.d/rocm-pin-600``` then install it with sudo apt update sudo apt install amdgpu-dkms sudo apt install rocm-hip-libraries
Thanks, I may have done something wrong but the rocminfo now works fine, I think the only thing left is changing the env variables.
what I found with the rocminfo command:
Agent 2
*******
Name: gfx1032
Uuid: GPU-XX
Marketing Name: AMD Radeon RX 6600M
Vendor Name: AMD
Feature: KERNEL_DISPATCH
Profile: BASE_PROFILE
Float Round Mode: NEAR
Max Queue Number: 128(0x80)
Queue Min Size: 64(0x40)
Queue Max Size: 131072(0x20000)
Queue Type: MULTI
Node: 1
Device Type: GPU
In that case you set your env variables as;
HSA_OVERRIDE_GFX_VERSION=10.3.0 HCC_AMDGPU_TARGET=gfx1032
I would advise you to do this in the terminal from where you are running your environment first by;
export HSA_OVERRIDE_GFX_VERSION=10.3.0
and export HCC_AMDGPU_TARGET=gfx1032
and then run the application.
If it works you can set the variables by adding them to your .bashrc
In that case you set your env variables as;
HSA_OVERRIDE_GFX_VERSION=10.3.0 HCC_AMDGPU_TARGET=gfx1032
I would advise you to do this in the terminal from where you are running your environment first by;
export HSA_OVERRIDE_GFX_VERSION=10.3.0
andexport HCC_AMDGPU_TARGET=gfx1032
and then run the application.If it works you can set the variables by adding them to your
.bashrc
I think everything went right but the error RuntimeError: No HIP GPUs are available still persists for me, but it was worth the try. Thanks for your time and let's wait to see if someone else can make it work
In that case you set your env variables as;
HSA_OVERRIDE_GFX_VERSION=10.3.0 HCC_AMDGPU_TARGET=gfx1032
I would advise you to do this in the terminal from where you are running your environment first by;
export HSA_OVERRIDE_GFX_VERSION=10.3.0
andexport HCC_AMDGPU_TARGET=gfx1032
and then run the application.If it works you can set the variables by adding them to your
.bashrc
I saw the script and tried out. this is what it printed out, a lot of erros
Checking ROCM support...
BAD: No ROCM devices found.
Checking PyTorch...
GOOD: PyTorch is working fine.
Checking user groups...
BAD: The user eduarte is NOT in RENDER and VIDEO groups. This is necessary in order to PyTorch use HIP resources
BAD: PyTorch ROCM support NOT found.
In that case add your user to those groups.
sudo usermod -a -G render eduarte
sudo usermod -a -G video eduarte
Check if the user was added.
sudo grep eduarte /etc/group
Then do a full reboot, not just a logout and login.
In that case add your user to those groups.
sudo usermod -a -G render eduarte
sudo usermod -a -G video eduarte
Check if the user was added.
sudo grep eduarte /etc/group
Then do a full reboot, not just a logout and login.
Omg it really worked, it took quite a time but it worked!! I think if this worked on my pc with an rx6600m it can pretty much work anywhere Thanks a lot for the help. These images took 5 minutes to generate but now I believe it's just low specs
In that case add your user to those groups.
sudo usermod -a -G render eduarte
sudo usermod -a -G video eduarte
Check if the user was added.
sudo grep eduarte /etc/group
Then do a full reboot, not just a logout and login.
Omg it really worked, it took quite a time but it worked!! I think if this worked on my pc with an rx6600m it can pretty much work anywhere Thanks a lot for the help. These images took 5 minutes to generate but now I believe it's just low specs
Glad to hear it, enjoy ✌️
I followed the steps provided by @ThomasMW , running into:
Speicherzugriffsfehler (Speicherabzug geschrieben)
which is "memoryacces failure" translated to english.
Then the task is being killed and the acces via localhost loses the connection.
My whole system also sometimes crashes when trying to python entry_with_update.py
, but not always.
It also crashed when I first did rocminfo
.
I did export HSA_OVERRIDE_GFX_VERSION=10.3.0
and export HCC_AMDGPU_TARGET=gfx1030
Also I tried the export commands bevore entering the venv and after, because I wasnt sure of the correct order.
When I run rocminfo outside the venv, agent2 name is gfx803 instead of 1030. I tried both, same error.
Would really apprecciate any help :)
Edit: I'm not sure if 10.3.0 is the correct gfx-version for me. I couldn't find any information about the correct version.
Edit2: Ok I tried following this post, but without the ComfyUI: https://github.com/comfyanonymous/ComfyUI/issues/650 Now it seems to do at least something, but loading the two new models still failes. I also tried the --use-split-cross-attention flag.
(venv) a@Zimt:~/fooocus/Fooocus$ python entry_with_update.py --preset realistic Already up-to-date Update succeeded. [System ARGV] ['entry_with_update.py', '--preset', 'realistic'] Loaded preset: /home/a/fooocus/Fooocus/presets/realistic.json Python 3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0] Fooocus version: 2.1.824 Running on local URL: http://127.0.0.1:7865
To create a public link, set share=True
in launch()
.
Total VRAM 8192 MB, total RAM 15904 MB
Set vram state to: NORMAL_VRAM
Disabling smart memory management
Device: cuda:0 AMD Radeon RX 590 Series : native
VAE dtype: torch.float32
Using sub quadratic optimization for cross attention, if you have memory or speed issues try using: --use-split-cross-attention
Refiner unloaded.
model_type EPS
adm 2816
Using split attention in VAE
Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
Using split attention in VAE
extra keys {'cond_stage_model.clip_l.text_projection', 'cond_stage_model.clip_g.transformer.text_model.embeddings.position_ids', 'cond_stage_model.clip_l.logit_scale'}
Base model loaded: /home/a/fooocus/Fooocus/models/checkpoints/realisticStockPhoto_v10.safetensors
Request to load LoRAs [['SDXL_FILM_PHOTOGRAPHY_STYLE_BetaV0.4.safetensors', 0.25], ['None', 1.0], ['None', 1.0], ['None', 1.0], ['None', 1.0]] for model [/home/a/fooocus/Fooocus/models/checkpoints/realisticStockPhoto_v10.safetensors].
Loaded LoRA [/home/a/fooocus/Fooocus/models/loras/SDXL_FILM_PHOTOGRAPHY_STYLE_BetaV0.4.safetensors] for UNet [/home/a/fooocus/Fooocus/models/checkpoints/realisticStockPhoto_v10.safetensors] with 788 keys at weight 0.25.
Loaded LoRA [/home/a/fooocus/Fooocus/models/loras/SDXL_FILM_PHOTOGRAPHY_STYLE_BetaV0.4.safetensors] for CLIP [/home/a/fooocus/Fooocus/models/checkpoints/realisticStockPhoto_v10.safetensors] with 264 keys at weight 0.25.
Fooocus V2 Expansion: Vocab with 642 words.
Fooocus Expansion engine loaded for cuda:0, use_fp16 = True.
Requested to load SDXLClipModel
Requested to load GPT2LMHeadModel
Loading 2 new models
Speicherzugriffsfehler (Speicherabzug geschrieben)
I think that I can help but first what's your system specification @wocif , and export HSA_OVERRIDE_GFX_VERSION=10.3.0 and export HCC_AMDGPU_TARGET=gfx1030 if based on your own gpu if your gpu is not an rx6600M the values is different
I think that I can help but first what's your system specification @wocif , and export HSA_OVERRIDE_GFX_VERSION=10.3.0 and export HCC_AMDGPU_TARGET=gfx1030 if based on your own gpu if your gpu is not an rx6600M the values is different
System: Linux Mint 21.2 Victoria base: Ubuntu 22.04 jammy CPU: AMD Ryzen 7 2700 GPU: AMD RX 590 8gb Python: 3.10.12 Environment: venv
I'm not certain about the HSA_OVERRIDE_GFX_VERSION, but as I tried to mentioned, the gpu actually is gfx803, but when I run rocminfo inside the venv it outputs me gfx1030. Should I use the output from inside or outside the venv? And how can I find the correct gfx-version for my gpu?
I think that I can help but first what's your system specification @wocif , and export HSA_OVERRIDE_GFX_VERSION=10.3.0 and export HCC_AMDGPU_TARGET=gfx1030 if based on your own gpu if your gpu is not an rx6600M the values is different
System: Linux Mint 21.2 Victoria base: Ubuntu 22.04 jammy CPU: AMD Ryzen 7 2700 GPU: AMD RX 590 8gb Python: 3.10.12 Environment: venv
I'm not certain about the HSA_OVERRIDE_GFX_VERSION, but as I tried to mentioned, the gpu actually is gfx803, but when I run rocminfo inside the venv it outputs me gfx1030. Should I use the output from inside or outside the venv? And how can I find the correct gfx-version for my gpu?
HCC_AMDGPU_TARGET=gfx803 based on https://www.techpowerup.com/gpu-specs/radeon-rx-590.c3322 HSA_OVERRIDE_GFX_VERSION=10.3.0 I think it's the same. and rocminfo technically you can use anywhere it was just a way to look for the gfx803 instead of techpowerup. but if rocminfo is not working the application won't work too
edit: rocminfo outputting gfx1030 something is going wrong I think but maybe not it could be outputting your cpu info? if I remember it outputs cpu and gpu info
My installation failed with a compilation error (I suspect an incompatibility with my kernel version, 6.5) and upon reboot my system was failing to boot GDM and so I only had a black screen with a blinking cursor. Took me quite a while in a TTY and several reboots until I got it booting correctly again.
For future reference...I had the same issue on a FRESH installation of OpenSuse TW. But fixed it: What I did from a fresh installation was:
In that case add your user to those groups.
sudo usermod -a -G render eduarte
sudo usermod -a -G video eduarte
Check if the user was added.
sudo grep eduarte /etc/group
Then do a full reboot, not just a logout and login.
I'm trying to run this inside a Docker-Container, but i cannot get it to work.
I got the same issues as you guys with telling me
RuntimeError: No HIP GPUs are available
So i tried your way installing ROCM: https://github.com/lllyasviel/Fooocus/issues/1079#issuecomment-1839145629 But this didnt went well for me,too. The repos aren't signed and i coulnd't get an updated repo-list, so i tried it this way:
On top of that it seems like there are some deleted links in one of your first workarounds https://github.com/lllyasviel/Fooocus/issues/1079#issuecomment-1837597622
Post-installation i added root to both user groups (problem here, i cannot restart docker container and also cannot implement it into the containerfile) but trying the rocminfo
command i ran into another issue:
Unable to open /dev/kfd read-write: Permission denied
So now i'm kinda clueless how to fix this since i cannot progress towards the next step..
P.S.: I'm running the container based on Ubuntu 22.04
I'm trying to run this inside a Docker-Container, but i cannot get it to work. I got the same issues as you guys with telling me
RuntimeError: No HIP GPUs are available
So i tried your way installing ROCM: https://github.com/lllyasviel/Fooocus/issues/1079#issuecomment-1839145629 But this didnt went well for me,too. The repos aren't signed and i coulnd't get an updated repo-list, so i tried it this way:
On top of that it seems like there are some deleted links in one of your first workarounds https://github.com/lllyasviel/Fooocus/issues/1079#issuecomment-1837597622
Post-installation i added root to both user groups (problem here, i cannot restart docker container and also cannot implement it into the containerfile) but trying the
rocminfo
command i ran into another issue:Unable to open /dev/kfd read-write: Permission denied
So now i'm kinda clueless how to fix this since i cannot progress towards the next step..P.S.: I'm running the container based on Ubuntu 22.04
I'm also clueless the moment, but will try it on docker too next month ig, since it's the better way to do stuff like that. Haven't had time to try yet
Confirming working with:
export HSA_OVERRIDE_GFX_VERSION=11.0.0
export HCC_AMDGPU_TARGET=gfx1101
python3 entry_with_update.py
on
Device: AMD Radeon RX 7800 XT (radeonsi, navi32, LLVM 17.0.6, DRM 3.54, 6.6.15-amd64) (0x747e)
Linux 6.6.15-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.6.15-2 (2024-02-04) x86_64 GNU/Linux
Torch packages built against rocm 5.7 not 5.6
Funnily enough, the GPU Agent doesn't show up on rocminfo
even when I set the envvars, but Fooocus itself works and is definitely using the GPU. Hope this helps someone.
From @led0nk on 2023-12-26:
The repos aren't signed and i coulnd't get an updated repo-list, so i tried it this way:
Yes, I confirm this. The ROCm installation guide is now at https://rocm.docs.amd.com/projects/install-on-linux/en/latest/how-to/native-install/ubuntu.html but when following the instructions, the repo does not have a public key signature.
W: GPG error: https://cloud.r-project.org/bin/linux/ubuntu jammy-cran40/ InRelease: The following signatures couldn't be verified because the public key is not available: NO_PUBKEY 51716619E084DAB9
E: The repository 'https://cloud.r-project.org/bin/linux/ubuntu jammy-cran40/ InRelease' is not signed.
N: Updating from such a repository can't be done securely, and is therefore disabled by default.
I fixed mine after updating my system and adding my user to the render
group.
RX 5700, Ubuntu 22.04.4
Many thanks to all the comments which helped me finding a solution!
doesnt work at all
haideraf@haideraf:~/Downloads$ sudo apt install amdgpu-dkms
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
amdgpu-dkms is already the newest version (1:6.7.0.60103-1787201.22.04).
0 upgraded, 0 newly installed, 0 to remove and 3 not upgraded.
haideraf@haideraf:~/Downloads$ sudo apt install rocm-hip-libraries
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
rocm-hip-libraries is already the newest version (6.1.3.60103-122~22.04).
rocm-hip-libraries set to manually installed.
0 upgraded, 0 newly installed, 0 to remove and 3 not upgraded.
haideraf@haideraf:~/Downloads$ rocminfo
ROCk module is NOT loaded, possibly no GPU devices
haideraf@haideraf:~/Downloads$ dmesg | grep amd
[ 0.000000] Linux version 6.5.0-41-generic (buildd@lcy02-amd64-120) (x86_64-linux-gnu-gcc-12 (Ubuntu 12.3.0-1ubuntu1~22.04) 12.3.0, GNU ld (GNU Binutils for Ubuntu) 2.38) #41~22.04.2-Ubuntu SMP PREEMPT_DYNAMIC Mon Jun 3 11:32:55 UTC 2 (Ubuntu 6.5.0-41.41~22.04.2-generic 6.5.13)
[ 0.652180] perf/amd_iommu: Detected AMD IOMMU #0 (2 banks, 4 counters/bank).
[ 0.754335] amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled
[ 6.024209] kvm_amd: TSC scaling supported
[ 6.024213] kvm_amd: Nested Virtualization enabled
[ 6.024214] kvm_amd: Nested Paging enabled
[ 6.024217] kvm_amd: SEV enabled (ASIDs 0 - 15)
[ 6.024219] kvm_amd: SEV-ES enabled (ASIDs 0 - 4294967295)
[ 6.024240] kvm_amd: Virtual VMLOAD VMSAVE supported
[ 6.024241] kvm_amd: Virtual GIF supported
[ 6.024242] kvm_amd: LBR virtualization supported
haideraf@haideraf:~/Downloads$ -^C
System: Ubuntu 22.04 CPU: i5-6400 GPU: RX 570 8G Environment: Anaconda Python: 3.10.13
I installed with the commands in
readme.md
(both Linux anaconda and Linux AMD)but it shows this error log:
I saw someone is discussing the same error log in ComfyUI and I tried what they said: downgrade Pytorch, ROCm to 5.4.2. but still got the same error.
EDIT: forgot to mention, the Web UI does appear, but the error message just print out after trying to generate things. I don't know if this information helps, but I'll leave it here.