lshqqytiger / stable-diffusion-webui-amdgpu-forge

Forge for stable-diffusion-webui-amdgpu (formerly stable-diffusion-webui-directml)
GNU Affero General Public License v3.0
45 stars 3 forks source link

RuntimeError: CUDA error: operation not supported ; Failed to load ZLUDA: list index out of range #36

Closed jdw-mis closed 5 days ago

jdw-mis commented 6 days ago

OS: Windows 11 Build 22631 GPU: 6700 XT Driver: 24.10.16 (Pro Edition) ROCm/HIP: both 5.7.1 and 6.1.2 Python: 3.10.6 I have already replaced the ROCm libraries for both ROCm versions with the gtx1031 libraries as detailed here: https://github.com/brknsoul/ROCmLibs/wiki/Installing-Forge%E2%80%90Zluda

WebUI launches, allows for full input, but when attempting to generate an image with any model it fails with the message:

RuntimeError: CUDA error: operation not supported
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.

Notably, prior in the log I've noticed the following lines:

fatal: No names found, cannot describe anything.
Failed to load ZLUDA: list index out of range
Using CPU-only torch
2024-09-21 16:59:18 DEBUG [PIL.Image] Image: failed to import FpxImagePlugin: No module named 'olefile'
2024-09-21 16:59:18 DEBUG [PIL.Image] Image: failed to import MicImagePlugin: No module named 'olefile'
ZLUDA device failed to pass basic operation test: index=0, device_name=AMD Radeon RX 6700 XT [ZLUDA]

I have attempted to install ZLUDA manually, it changed nothing. Same with deleting one or the other ROCm version paths. I think its just not loading ZLUDA for some incomprehensible reason.

Full Log ``` venv "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\Scripts\Python.exe" WARNING: ZLUDA works best with SD.Next. Please consider migrating to SD.Next. fatal: No names found, cannot describe anything. Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: f2.0.1v1.10.1-1.10.1 Commit hash: 2e4a9d6f849294b4f2988d8e7f93bf10259e52cb Using ZLUDA in C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\.zluda Failed to load ZLUDA: list index out of range Using CPU-only torch 2024-09-21 16:58:58 DEBUG [root] Installing put extensions here.txt 2024-09-21 16:58:58 DEBUG [root] Installing extra-options-section 2024-09-21 16:58:58 DEBUG [root] Installing forge_legacy_preprocessors 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_inpaint 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_marigold 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_normalbae 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_recolor 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_reference 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_revision 2024-09-21 16:58:58 DEBUG [root] Installing forge_preprocessor_tile 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_animagine_xl_31 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_birefnet 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_example 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_florence_2 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_geowizard 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_iclight 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_idm_vton 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_illusion_diffusion 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_photo_maker_v2 2024-09-21 16:58:58 DEBUG [root] Installing forge_space_sapiens_normal 2024-09-21 16:58:58 DEBUG [root] Installing mobile 2024-09-21 16:58:58 DEBUG [root] Installing prompt-bracket-checker 2024-09-21 16:58:58 DEBUG [root] Installing ScuNET 2024-09-21 16:58:58 DEBUG [root] Installing sd_forge_controlllite 2024-09-21 16:58:58 DEBUG [root] Installing sd_forge_controlnet 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_dynamic_thresholding 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_fooocus_inpaint 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_freeu 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_ipadapter 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_kohya_hrfix 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_latent_modifier 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_lora 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_multidiffusion 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_neveroom 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_perturbed_attention 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_sag 2024-09-21 16:58:59 DEBUG [root] Installing sd_forge_stylealign 2024-09-21 16:58:59 DEBUG [root] Installing soft-inpainting 2024-09-21 16:58:59 DEBUG [root] Installing SwinIR Launching Web UI with arguments: --theme dark --use-zluda --loglevel debug Total VRAM 12272 MB, total RAM 32621 MB pytorch version: 2.3.1+cu121 Set vram state to: NORMAL_VRAM Device: cuda:0 AMD Radeon RX 6700 XT [ZLUDA] : native VAE dtype preferences: [torch.bfloat16, torch.float32] -> torch.bfloat16 CUDA Using Stream: False 2024-09-21 16:59:11 DEBUG [httpx] load_ssl_context verify=True cert=None trust_env=True http2=False 2024-09-21 16:59:11 DEBUG [httpx] load_verify_locations cafile='C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\venv\\lib\\site-packages\\certifi\\cacert.pem' 2024-09-21 16:59:11 DEBUG [httpx] load_ssl_context verify=True cert=None trust_env=True http2=False 2024-09-21 16:59:11 DEBUG [httpx] load_verify_locations cafile='C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\venv\\lib\\site-packages\\certifi\\cacert.pem' 2024-09-21 16:59:11 DEBUG [httpx] load_ssl_context verify=True cert=None trust_env=True http2=False 2024-09-21 16:59:11 DEBUG [httpx] load_verify_locations cafile='C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\venv\\lib\\site-packages\\certifi\\cacert.pem' 2024-09-21 16:59:11 DEBUG [bitsandbytes.cextension] Loading bitsandbytes native library from: C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cuda121.dll Using pytorch cross attention Using pytorch attention for VAE 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing BlpImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing BmpImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing BufrStubImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing CurImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing DcxImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing DdsImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing EpsImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing FitsImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing FitsStubImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing FliImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing FpxImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Image: failed to import FpxImagePlugin: No module named 'olefile' 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing FtexImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing GbrImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing GifImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing GribStubImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing Hdf5StubImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing IcnsImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing IcoImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing ImImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing ImtImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing IptcImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing JpegImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing Jpeg2KImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing McIdasImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing MicImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Image: failed to import MicImagePlugin: No module named 'olefile' 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing MpegImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing MpoImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing MspImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PalmImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PcdImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PcxImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PdfImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PixarImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PngImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PpmImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing PsdImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing QoiImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing SgiImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing SpiderImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing SunImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing TgaImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing TiffImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing WebPImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing WmfImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing XbmImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing XpmImagePlugin 2024-09-21 16:59:18 DEBUG [PIL.Image] Importing XVThumbImagePlugin 2024-09-21 16:59:23 DEBUG [matplotlib] matplotlib data path: C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\matplotlib\mpl-data 2024-09-21 16:59:23 DEBUG [matplotlib] CONFIGDIR=C:\Users\MEDIUM_MK_5_0\.matplotlib 2024-09-21 16:59:23 DEBUG [matplotlib] interactive is False 2024-09-21 16:59:23 DEBUG [matplotlib] platform is win32 2024-09-21 16:59:23 DEBUG [matplotlib] CACHEDIR=C:\Users\MEDIUM_MK_5_0\.matplotlib 2024-09-21 16:59:23 DEBUG [matplotlib.font_manager] Using fontManager instance from C:\Users\MEDIUM_MK_5_0\.matplotlib\fontlist-v390.json 2024-09-21 16:59:24 DEBUG [git.cmd] Popen(['git', 'version'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) 2024-09-21 16:59:24 DEBUG [git.cmd] Popen(['git', 'version'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) ONNX: version=1.19.2 provider=CPUExecutionProvider, available=['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider'] ZLUDA device failed to pass basic operation test: index=0, device_name=AMD Radeon RX 6700 XT [ZLUDA] CUDA error: operation not supported CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. ControlNet preprocessor location: C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\models\ControlNetPreprocessor 2024-09-21 16:59:30 DEBUG [asyncio] Using selector: SelectSelector 2024-09-21 16:59:31,709 - ControlNet - INFO - ControlNet UI callback registered. Model selected: {'checkpoint_info': {'filename': 'C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\models\\Stable-diffusion\\AbyssOrangeMix2_nsfw.safetensors', 'hash': 'a87fd7da'}, 'additional_modules': [], 'unet_storage_dtype': None} Using online LoRAs in FP16: False 2024-09-21 16:59:33 DEBUG [asyncio] Using selector: SelectSelector Running on local URL: http://127.0.0.1:7860 2024-09-21 16:59:34 DEBUG [httpx] load_ssl_context verify=None cert=None trust_env=True http2=False 2024-09-21 16:59:34 INFO [httpx] HTTP Request: GET http://127.0.0.1:7860/startup-events "HTTP/1.1 200 OK" 2024-09-21 16:59:34 DEBUG [httpx] load_ssl_context verify=False cert=None trust_env=True http2=False 2024-09-21 16:59:34 INFO [httpx] HTTP Request: HEAD http://127.0.0.1:7860/ "HTTP/1.1 200 OK" To create a public link, set `share=True` in `launch()`. Startup time: 39.6s (prepare environment: 4.5s, import torch: 25.3s, initialize shared: 2.7s, other imports: 0.1s, load scripts: 2.8s, create ui: 2.8s, gradio launch: 1.6s). 2024-09-21 16:59:38 DEBUG [matplotlib.pyplot] Loaded backend tkagg version 8.6. 2024-09-21 16:59:38 DEBUG [matplotlib.pyplot] Loaded backend agg version v2.2. 2024-09-21 16:59:38 DEBUG [matplotlib.pyplot] Loaded backend tkagg version 8.6. 2024-09-21 16:59:38 DEBUG [matplotlib.pyplot] Loaded backend agg version v2.2. 2024-09-21 16:59:38 DEBUG [matplotlib.pyplot] Loaded backend tkagg version 8.6. 2024-09-21 16:59:38 DEBUG [matplotlib.pyplot] Loaded backend agg version v2.2. Environment vars changed: {'stream': False, 'inference_memory': 1024.0, 'pin_shared_memory': False} [GPU Setting] You will use 91.66% GPU memory (11248.00 MB) to load weights, and use 8.34% GPU memory (1024.00 MB) to do matrix computation. 2024-09-21 16:59:38 DEBUG [git.cmd] Popen(['git', 'remote', 'get-url', '--all', 'origin'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) 2024-09-21 16:59:38 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch-check'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 16:59:38 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 16:59:38 DEBUG [git.cmd] Popen(['git', 'remote', 'get-url', '--all', 'origin'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) 2024-09-21 16:59:39 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch-check'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 16:59:39 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_begin with no data 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_header_field with data[42:61] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_header_value with data[63:119] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_header_end with no data 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_header_field with data[121:133] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_header_value with data[135:144] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_header_end with no data 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_headers_finished with no data 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[148:152] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:2] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[154:32768] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:83824] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:1] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[83825:262144] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:34493] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:1] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[34494:36008] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:1] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[36009:64741] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:1] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_data with data[64742:244716] 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_part_end with no data 2024-09-21 17:00:19 DEBUG [multipart.multipart] Calling on_end with no data 2024-09-21 17:00:19 DEBUG [PIL.PngImagePlugin] STREAM b'IHDR' 16 13 2024-09-21 17:00:19 DEBUG [PIL.PngImagePlugin] STREAM b'tEXt' 41 873 2024-09-21 17:00:19 DEBUG [PIL.PngImagePlugin] STREAM b'IDAT' 926 65536 2024-09-21 17:00:21 INFO [modules.shared_state] Starting job task(7wz5ssxykbpqt06) Exception in thread MemMon: Traceback (most recent call last): File "C:\Program Files\Python310\lib\threading.py", line 1016, in _bootstrap_inner self.run() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\memmon.py", line 43, in run torch.cuda.reset_peak_memory_stats() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\cuda\memory.py", line 309, in reset_peak_memory_stats return torch._C._cuda_resetPeakMemoryStats(device) RuntimeError: invalid argument to reset_peak_memory_stats Loading Model: {'checkpoint_info': {'filename': 'C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\models\\Stable-diffusion\\AbyssOrangeMix2_nsfw.safetensors', 'hash': 'a87fd7da'}, 'additional_modules': [], 'unet_storage_dtype': None} [Unload] Trying to free all memory for cuda:0 with 0 models keep loaded ... Done. StateDict Keys: {'unet': 686, 'vae': 248, 'text_encoder': 197, 'ignore': 0} C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\transformers\tokenization_utils_base.py:1601: FutureWarning: `clean_up_tokenization_spaces` was not set. It will be set to `True` by default. This behavior will be depracted in transformers v4.45, and will be then set to `False` by default. For more details check this issue: https://github.com/huggingface/transformers/issues/31884 warnings.warn( Working with z of shape (1, 4, 32, 32) = 4096 dimensions. K-Model Created: {'storage_dtype': torch.float16, 'computation_dtype': torch.float16} Model loaded in 1.4s (unload existing model: 0.2s, forge model load: 1.2s). [Unload] Trying to free 1329.14 MB for cuda:0 with 0 models keep loaded ... Done. [Memory Management] Target: JointTextEncoder, Free GPU: 10704.50 MB, Model Require: 234.72 MB, Previously Loaded: 0.00 MB, Inference Require: 1024.00 MB, Remaining: 9445.78 MB, All loaded to GPU. Moving model(s) has taken 0.40 seconds Traceback (most recent call last): File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules_forge\main_thread.py", line 30, in work self.result = self.func(*self.args, **self.kwargs) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\txt2img.py", line 123, in txt2img_function processed = processing.process_images(p) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 818, in process_images res = process_images_inner(p) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 1023, in process_images_inner p.setup_conds() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 1619, in setup_conds super().setup_conds() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 501, in setup_conds self.uc = self.get_conds_with_caching(prompt_parser.get_learned_conditioning, negative_prompts, total_steps, [self.cached_uc], self.extra_network_data) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 472, in get_conds_with_caching cache[1] = function(shared.sd_model, required_prompts, steps, hires_steps, shared.opts.use_old_scheduling) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\prompt_parser.py", line 189, in get_learned_conditioning conds = model.get_learned_conditioning(texts) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context return func(*args, **kwargs) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\diffusion_engine\sd15.py", line 63, in get_learned_conditioning cond = self.text_processing_engine(prompt) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\text_processing\classic_engine.py", line 268, in __call__ z = self.process_tokens(tokens, multipliers) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\text_processing\classic_engine.py", line 301, in process_tokens z = self.encode_with_transformers(tokens) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\text_processing\classic_engine.py", line 126, in encode_with_transformers self.text_encoder.transformer.text_model.embeddings.position_embedding = self.text_encoder.transformer.text_model.embeddings.position_embedding.to(dtype=torch.float32) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\nn\modules\module.py", line 1173, in to return self._apply(convert) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\nn\modules\module.py", line 804, in _apply param_applied = fn(param) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\nn\modules\module.py", line 1159, in convert return t.to( RuntimeError: CUDA error: operation not supported CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. CUDA error: operation not supported CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2024-09-21 17:00:23 INFO [modules.shared_state] Ending job task(7wz5ssxykbpqt06) (1.91 seconds) ```
TheFerumn commented 6 days ago

Definitely Zluda error. Try to uninstall HIP SDK 6.1.2, reinstall HIP SDK 5.7.1 put optimized rocmlibs again and make sure you put in correct folder. Then before launching Forge delete Venv folder. If it won't work you might try to delete venv folder again and change command into --zluda i am not sure if --use-zluda still works correctly but i was installing only on --zluda command without any issues

jdw-mis commented 6 days ago

Uninstalled both ROCm, made sure Program Files folder was clear, installed 5.7.1, backed up original library files, overwrote with gfx1031 library ontop, deleted the entire sd forge folder and recloned from git, put AOM2 in models folder, and changed launch param to --zluda --loglevel debug even though --use-zluda should be a valid alias. Same errors as before.

Log ``` Creating venv in directory C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv using python "C:\Program Files\Python310\python.exe" Requirement already satisfied: pip in c:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages (22.2.1) Collecting pip Using cached pip-24.2-py3-none-any.whl (1.8 MB) Installing collected packages: pip Attempting uninstall: pip Found existing installation: pip 22.2.1 Uninstalling pip-22.2.1: Successfully uninstalled pip-22.2.1 Successfully installed pip-24.2 venv "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\Scripts\Python.exe" ROCm Toolkit 5.7 was found. fatal: No names found, cannot describe anything. Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: f2.0.1v1.10.1-1.10.1 Commit hash: 2e4a9d6f849294b4f2988d8e7f93bf10259e52cb Using ZLUDA in C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\.zluda Failed to load ZLUDA: list index out of range Using CPU-only torch Installing torch and torchvision Looking in indexes: https://pypi.org/simple, https://download.pytorch.org/whl/cu121 Collecting torch==2.3.1 Using cached https://download.pytorch.org/whl/cu121/torch-2.3.1%2Bcu121-cp310-cp310-win_amd64.whl (2423.5 MB) Collecting torchvision Using cached https://download.pytorch.org/whl/cu121/torchvision-0.19.1%2Bcu121-cp310-cp310-win_amd64.whl (5.8 MB) Collecting filelock (from torch==2.3.1) Using cached filelock-3.16.1-py3-none-any.whl.metadata (2.9 kB) Collecting typing-extensions>=4.8.0 (from torch==2.3.1) Using cached typing_extensions-4.12.2-py3-none-any.whl.metadata (3.0 kB) Collecting sympy (from torch==2.3.1) Using cached sympy-1.13.3-py3-none-any.whl.metadata (12 kB) Collecting networkx (from torch==2.3.1) Using cached networkx-3.3-py3-none-any.whl.metadata (5.1 kB) Collecting jinja2 (from torch==2.3.1) Using cached jinja2-3.1.4-py3-none-any.whl.metadata (2.6 kB) Collecting fsspec (from torch==2.3.1) Using cached fsspec-2024.9.0-py3-none-any.whl.metadata (11 kB) Collecting mkl<=2021.4.0,>=2021.1.1 (from torch==2.3.1) Using cached https://download.pytorch.org/whl/mkl-2021.4.0-py2.py3-none-win_amd64.whl (228.5 MB) Collecting numpy (from torchvision) Using cached numpy-2.1.1-cp310-cp310-win_amd64.whl.metadata (59 kB) INFO: pip is looking at multiple versions of torchvision to determine which version is compatible with other requirements. This could take a while. Collecting torchvision Using cached torchvision-0.19.1-cp310-cp310-win_amd64.whl.metadata (6.1 kB) Using cached https://download.pytorch.org/whl/cu121/torchvision-0.19.0%2Bcu121-cp310-cp310-win_amd64.whl (5.8 MB) Using cached torchvision-0.19.0-1-cp310-cp310-win_amd64.whl.metadata (6.1 kB) Collecting numpy<2 (from torchvision) Using cached numpy-1.26.4-cp310-cp310-win_amd64.whl.metadata (61 kB) Collecting torchvision Using cached https://download.pytorch.org/whl/cu121/torchvision-0.18.1%2Bcu121-cp310-cp310-win_amd64.whl (5.7 MB) Collecting pillow!=8.3.*,>=5.3.0 (from torchvision) Using cached pillow-10.4.0-cp310-cp310-win_amd64.whl.metadata (9.3 kB) Collecting intel-openmp==2021.* (from mkl<=2021.4.0,>=2021.1.1->torch==2.3.1) Using cached https://download.pytorch.org/whl/intel_openmp-2021.4.0-py2.py3-none-win_amd64.whl (3.5 MB) Collecting tbb==2021.* (from mkl<=2021.4.0,>=2021.1.1->torch==2.3.1) Using cached tbb-2021.13.1-py3-none-win_amd64.whl.metadata (1.1 kB) Collecting MarkupSafe>=2.0 (from jinja2->torch==2.3.1) Using cached https://download.pytorch.org/whl/MarkupSafe-2.1.5-cp310-cp310-win_amd64.whl (17 kB) Collecting mpmath<1.4,>=1.1.0 (from sympy->torch==2.3.1) Using cached https://download.pytorch.org/whl/mpmath-1.3.0-py3-none-any.whl (536 kB) Using cached tbb-2021.13.1-py3-none-win_amd64.whl (286 kB) Using cached pillow-10.4.0-cp310-cp310-win_amd64.whl (2.6 MB) Using cached typing_extensions-4.12.2-py3-none-any.whl (37 kB) Using cached filelock-3.16.1-py3-none-any.whl (16 kB) Using cached fsspec-2024.9.0-py3-none-any.whl (179 kB) Using cached jinja2-3.1.4-py3-none-any.whl (133 kB) Using cached networkx-3.3-py3-none-any.whl (1.7 MB) Using cached numpy-2.1.1-cp310-cp310-win_amd64.whl (12.9 MB) Using cached sympy-1.13.3-py3-none-any.whl (6.2 MB) Installing collected packages: tbb, mpmath, intel-openmp, typing-extensions, sympy, pillow, numpy, networkx, mkl, MarkupSafe, fsspec, filelock, jinja2, torch, torchvision Successfully installed MarkupSafe-2.1.5 filelock-3.16.1 fsspec-2024.9.0 intel-openmp-2021.4.0 jinja2-3.1.4 mkl-2021.4.0 mpmath-1.3.0 networkx-3.3 numpy-2.1.1 pillow-10.4.0 sympy-1.13.3 tbb-2021.13.1 torch-2.3.1+cu121 torchvision-0.18.1+cu121 typing-extensions-4.12.2 Installing clip Installing open_clip Cloning assets into C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\repositories\stable-diffusion-webui-assets... Cloning into 'C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\repositories\stable-diffusion-webui-assets'... remote: Enumerating objects: 20, done. remote: Counting objects: 100% (20/20), done. remote: Compressing objects: 100% (18/18), done. Receiving objects: 100% (20/20), 132.70 KiB | 26.54 MiB/s, done.from 0)Receiving objects: 100% (20/20) Cloning huggingface_guess into C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\repositories\huggingface_guess... Cloning into 'C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\repositories\huggingface_guess'... remote: Enumerating objects: 46, done. remote: Counting objects: 100% (46/46), done. remote: Compressing objects: 100% (33/33), done. remote: Total 46 (delta 26), reused 33 (delta 13), pack-reused 0 (from 0) Receiving objects: 100% (46/46), 35.63 KiB | 7.13 MiB/s, done. Resolving deltas: 100% (26/26), done. Cloning BLIP into C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\repositories\BLIP... Cloning into 'C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\repositories\BLIP'... remote: Enumerating objects: 277, done. remote: Counting objects: 100% (183/183), done. remote: Compressing objects: 100% (46/46), done. remote: Total 277 (delta 145), reused 137 (delta 137), pack-reused 94 (from 1) Receiving objects: 100% (277/277), 7.04 MiB | 27.71 MiB/s, done. Resolving deltas: 100% (152/152), done. Installing requirements Installing onnxruntime-gpu 2024-09-21 19:07:42 DEBUG [root] Installing put extensions here.txt 2024-09-21 19:07:42 DEBUG [root] Installing extra-options-section 2024-09-21 19:07:42 DEBUG [root] Installing forge_legacy_preprocessors Installing forge_legacy_preprocessor requirement: fvcore Installing forge_legacy_preprocessor requirement: mediapipe Installing forge_legacy_preprocessor requirement: svglib Installing forge_legacy_preprocessor requirement: insightface Installing forge_legacy_preprocessor requirement: handrefinerportable Installing forge_legacy_preprocessor requirement: depth_anything Installing forge_legacy_preprocessor requirement: depth_anything_v2 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_inpaint 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_marigold 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_normalbae 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_recolor 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_reference 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_revision 2024-09-21 19:09:49 DEBUG [root] Installing forge_preprocessor_tile 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_animagine_xl_31 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_birefnet 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_example 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_florence_2 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_geowizard 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_iclight 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_idm_vton 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_illusion_diffusion 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_photo_maker_v2 2024-09-21 19:09:49 DEBUG [root] Installing forge_space_sapiens_normal 2024-09-21 19:09:49 DEBUG [root] Installing mobile 2024-09-21 19:09:49 DEBUG [root] Installing prompt-bracket-checker 2024-09-21 19:09:49 DEBUG [root] Installing ScuNET 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_controlllite 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_controlnet 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_dynamic_thresholding 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_fooocus_inpaint 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_freeu 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_ipadapter 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_kohya_hrfix 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_latent_modifier 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_lora 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_multidiffusion 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_neveroom 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_perturbed_attention 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_sag 2024-09-21 19:09:49 DEBUG [root] Installing sd_forge_stylealign 2024-09-21 19:09:49 DEBUG [root] Installing soft-inpainting 2024-09-21 19:09:49 DEBUG [root] Installing SwinIR Launching Web UI with arguments: --zluda --loglevel debug Total VRAM 12272 MB, total RAM 32621 MB pytorch version: 2.3.1+cu121 Set vram state to: NORMAL_VRAM Device: cuda:0 AMD Radeon RX 6700 XT [ZLUDA] : native VAE dtype preferences: [torch.bfloat16, torch.float32] -> torch.bfloat16 Installing bitsandbytes==0.43.3 CUDA Using Stream: False 2024-09-21 19:10:07 DEBUG [httpx] load_ssl_context verify=True cert=None trust_env=True http2=False 2024-09-21 19:10:07 DEBUG [httpx] load_verify_locations cafile='C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\venv\\lib\\site-packages\\certifi\\cacert.pem' 2024-09-21 19:10:07 DEBUG [httpx] load_ssl_context verify=True cert=None trust_env=True http2=False 2024-09-21 19:10:07 DEBUG [httpx] load_verify_locations cafile='C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\venv\\lib\\site-packages\\certifi\\cacert.pem' 2024-09-21 19:10:07 DEBUG [httpx] load_ssl_context verify=True cert=None trust_env=True http2=False 2024-09-21 19:10:07 DEBUG [httpx] load_verify_locations cafile='C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\venv\\lib\\site-packages\\certifi\\cacert.pem' 2024-09-21 19:10:07 DEBUG [bitsandbytes.cextension] Loading bitsandbytes native library from: C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cuda121.dll The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`. 0it [00:00, ?it/s] Using pytorch cross attention Using pytorch attention for VAE 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing BlpImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing BmpImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing BufrStubImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing CurImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing DcxImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing DdsImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing EpsImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing FitsImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing FitsStubImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing FliImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing FpxImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Image: failed to import FpxImagePlugin: No module named 'olefile' 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing FtexImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing GbrImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing GifImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing GribStubImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing Hdf5StubImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing IcnsImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing IcoImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing ImImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing ImtImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing IptcImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing JpegImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing Jpeg2KImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing McIdasImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing MicImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Image: failed to import MicImagePlugin: No module named 'olefile' 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing MpegImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing MpoImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing MspImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PalmImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PcdImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PcxImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PdfImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PixarImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PngImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PpmImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing PsdImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing QoiImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing SgiImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing SpiderImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing SunImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing TgaImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing TiffImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing WebPImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing WmfImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing XbmImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing XpmImagePlugin 2024-09-21 19:10:20 DEBUG [PIL.Image] Importing XVThumbImagePlugin 2024-09-21 19:10:25 DEBUG [matplotlib] matplotlib data path: C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\matplotlib\mpl-data 2024-09-21 19:10:25 DEBUG [matplotlib] CONFIGDIR=C:\Users\MEDIUM_MK_5_0\.matplotlib 2024-09-21 19:10:25 DEBUG [matplotlib] interactive is False 2024-09-21 19:10:25 DEBUG [matplotlib] platform is win32 2024-09-21 19:10:25 DEBUG [matplotlib] CACHEDIR=C:\Users\MEDIUM_MK_5_0\.matplotlib 2024-09-21 19:10:25 DEBUG [matplotlib.font_manager] Using fontManager instance from C:\Users\MEDIUM_MK_5_0\.matplotlib\fontlist-v390.json 2024-09-21 19:10:32 DEBUG [git.cmd] Popen(['git', 'version'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) 2024-09-21 19:10:32 DEBUG [git.cmd] Popen(['git', 'version'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) ONNX: version=1.19.2 provider=CPUExecutionProvider, available=['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider'] ZLUDA device failed to pass basic operation test: index=0, device_name=AMD Radeon RX 6700 XT [ZLUDA] CUDA error: operation not supported CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. ControlNet preprocessor location: C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\models\ControlNetPreprocessor 2024-09-21 19:10:38 DEBUG [asyncio] Using selector: SelectSelector 2024-09-21 19:10:39,594 - ControlNet - INFO - ControlNet UI callback registered. Model selected: {'checkpoint_info': {'filename': 'C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\models\\Stable-diffusion\\AbyssOrangeMix2_nsfw.safetensors', 'hash': 'a87fd7da'}, 'additional_modules': [], 'unet_storage_dtype': None} Using online LoRAs in FP16: False 2024-09-21 19:10:41 DEBUG [asyncio] Using selector: SelectSelector Running on local URL: http://127.0.0.1:7860 2024-09-21 19:10:41 DEBUG [httpx] load_ssl_context verify=None cert=None trust_env=True http2=False 2024-09-21 19:10:41 INFO [httpx] HTTP Request: GET http://127.0.0.1:7860/startup-events "HTTP/1.1 200 OK" 2024-09-21 19:10:41 DEBUG [httpx] load_ssl_context verify=False cert=None trust_env=True http2=False 2024-09-21 19:10:42 INFO [httpx] HTTP Request: HEAD http://127.0.0.1:7860/ "HTTP/1.1 200 OK" To create a public link, set `share=True` in `launch()`. Startup time: 371.1s (prepare environment: 325.9s, import torch: 36.3s, setup paths: 6.4s, initialize shared: 3.0s, other imports: 0.1s, list SD models: 0.1s, load scripts: 2.8s, create ui: 2.9s, gradio launch: 1.4s). 2024-09-21 19:10:45 DEBUG [matplotlib.pyplot] Loaded backend tkagg version 8.6. 2024-09-21 19:10:45 DEBUG [matplotlib.pyplot] Loaded backend agg version v2.2. 2024-09-21 19:10:45 DEBUG [matplotlib.pyplot] Loaded backend tkagg version 8.6. 2024-09-21 19:10:45 DEBUG [matplotlib.pyplot] Loaded backend agg version v2.2. 2024-09-21 19:10:45 DEBUG [matplotlib.pyplot] Loaded backend tkagg version 8.6. 2024-09-21 19:10:45 DEBUG [matplotlib.pyplot] Loaded backend agg version v2.2. Environment vars changed: {'stream': False, 'inference_memory': 1024.0, 'pin_shared_memory': False} [GPU Setting] You will use 91.66% GPU memory (11248.00 MB) to load weights, and use 8.34% GPU memory (1024.00 MB) to do matrix computation. 2024-09-21 19:10:45 DEBUG [git.cmd] Popen(['git', 'remote', 'get-url', '--all', 'origin'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) 2024-09-21 19:10:45 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch-check'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 19:10:45 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 19:10:45 DEBUG [git.cmd] Popen(['git', 'remote', 'get-url', '--all', 'origin'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=None) 2024-09-21 19:10:46 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch-check'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 19:10:46 DEBUG [git.cmd] Popen(['git', 'cat-file', '--batch'], cwd=C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge, universal_newlines=False, shell=None, istream=) 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_begin with no data 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_header_field with data[42:61] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_header_value with data[63:119] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_header_end with no data 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_header_field with data[121:133] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_header_value with data[135:144] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_header_end with no data 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_headers_finished with no data 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[148:152] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:2] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[154:65536] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:96726] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:1] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[96727:262144] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_data with data[0:211948] 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_part_end with no data 2024-09-21 19:11:19 DEBUG [multipart.multipart] Calling on_end with no data 2024-09-21 19:11:19 DEBUG [PIL.PngImagePlugin] STREAM b'IHDR' 16 13 2024-09-21 19:11:19 DEBUG [PIL.PngImagePlugin] STREAM b'tEXt' 41 873 2024-09-21 19:11:19 DEBUG [PIL.PngImagePlugin] STREAM b'IDAT' 926 65536 2024-09-21 19:11:25 INFO [modules.shared_state] Starting job task(m0j4lm0ihzufmbv) Exception in thread MemMon: Traceback (most recent call last): File "C:\Program Files\Python310\lib\threading.py", line 1016, in _bootstrap_inner self.run() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\memmon.py", line 43, in run torch.cuda.reset_peak_memory_stats() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\cuda\memory.py", line 309, in reset_peak_memory_stats return torch._C._cuda_resetPeakMemoryStats(device) Loading Model: {'checkpoint_info': {'filename': 'C:\\stable-diffusion-webui\\stable-diffusion-webui-amdgpu-forge\\models\\Stable-diffusion\\AbyssOrangeMix2_nsfw.safetensors', 'hash': 'a87fd7da'}, 'additional_modules': [], 'unet_storage_dtype': None} RuntimeError: invalid argument to reset_peak_memory_stats [Unload] Trying to free all memory for cuda:0 with 0 models keep loaded ... Done. StateDict Keys: {'unet': 686, 'vae': 248, 'text_encoder': 197, 'ignore': 0} C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\transformers\tokenization_utils_base.py:1601: FutureWarning: `clean_up_tokenization_spaces` was not set. It will be set to `True` by default. This behavior will be depracted in transformers v4.45, and will be then set to `False` by default. For more details check this issue: https://github.com/huggingface/transformers/issues/31884 warnings.warn( Working with z of shape (1, 4, 32, 32) = 4096 dimensions. K-Model Created: {'storage_dtype': torch.float16, 'computation_dtype': torch.float16} Calculating sha256 for C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\models\Stable-diffusion\AbyssOrangeMix2_nsfw.safetensors: 9292844753aaaf24ea39412c0159063a4e620afef2ca838416f856e52da22cd2 Model loaded in 1.6s (unload existing model: 0.2s, forge model load: 1.4s). [Unload] Trying to free 1329.14 MB for cuda:0 with 0 models keep loaded ... Done. [Memory Management] Target: JointTextEncoder, Free GPU: 11954.00 MB, Model Require: 234.72 MB, Previously Loaded: 0.00 MB, Inference Require: 1024.00 MB, Remaining: 10695.28 MB, All loaded to GPU. Moving model(s) has taken 0.42 seconds Traceback (most recent call last): File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules_forge\main_thread.py", line 30, in work self.result = self.func(*self.args, **self.kwargs) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\txt2img.py", line 123, in txt2img_function processed = processing.process_images(p) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 818, in process_images res = process_images_inner(p) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 1023, in process_images_inner p.setup_conds() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 1619, in setup_conds super().setup_conds() File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 501, in setup_conds self.uc = self.get_conds_with_caching(prompt_parser.get_learned_conditioning, negative_prompts, total_steps, [self.cached_uc], self.extra_network_data) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\processing.py", line 472, in get_conds_with_caching cache[1] = function(shared.sd_model, required_prompts, steps, hires_steps, shared.opts.use_old_scheduling) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\modules\prompt_parser.py", line 189, in get_learned_conditioning conds = model.get_learned_conditioning(texts) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context return func(*args, **kwargs) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\diffusion_engine\sd15.py", line 63, in get_learned_conditioning cond = self.text_processing_engine(prompt) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\text_processing\classic_engine.py", line 268, in __call__ z = self.process_tokens(tokens, multipliers) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\text_processing\classic_engine.py", line 301, in process_tokens z = self.encode_with_transformers(tokens) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\backend\text_processing\classic_engine.py", line 126, in encode_with_transformers self.text_encoder.transformer.text_model.embeddings.position_embedding = self.text_encoder.transformer.text_model.embeddings.position_embedding.to(dtype=torch.float32) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\nn\modules\module.py", line 1173, in to return self._apply(convert) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\nn\modules\module.py", line 804, in _apply param_applied = fn(param) File "C:\stable-diffusion-webui\stable-diffusion-webui-amdgpu-forge\venv\lib\site-packages\torch\nn\modules\module.py", line 1159, in convert return t.to( RuntimeError: CUDA error: operation not supported CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. CUDA error: operation not supported CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2024-09-21 19:11:27 INFO [modules.shared_state] Ending job task(m0j4lm0ihzufmbv) (2.07 seconds) ```
jdw-mis commented 5 days ago

I fixed it by manually adding the generated .zluda folder to Path then rebuilding the venv. Bizarre.

TheFerumn commented 5 days ago

I was going to ask if your folder is added to PATH but i was like "naaah it was added automatically when i was testing zluda". Kind of funny how on 1 PC it adds itself but on other it doesn't. Windows is like a lottery :D