Open LYC878484 opened 8 months ago
You can't skip cuda test with ZLUDA, and also you can't use the virtual environment that used with DirectML. Remove or rename (if you want to backup) venv folder and try again.
You can't skip cuda test with ZLUDA, and also you can't use the virtual environment that used with DirectML. Remove or rename (if you want to backup) venv folder and try again.
follow you suggestion, rename venv to venv_directML and process .\webui-user.bat again, error:Failed to automatically patch torch with ZLUDA. Could not find ZLUDA from PATH.
now set COMMANDLINE_ARGS=--use-zluda --opt-sub-quad-attention --lowvram --disable-nan-check
Collecting mpmath>=0.19 Using cached https://download.pytorch.org/whl/mpmath-1.3.0-py3-none-any.whl (536 kB) Installing collected packages: mpmath, urllib3, typing-extensions, sympy, pillow, numpy, networkx, MarkupSafe, idna, fsspec, filelock, charset-normalizer, certifi, requests, jinja2, torch, torchvision Successfully installed MarkupSafe-2.1.3 certifi-2022.12.7 charset-normalizer-2.1.1 filelock-3.9.0 fsspec-2023.4.0 idna-3.4 jinja2-3.1.2 mpmath-1.3.0 networkx-3.2.1 numpy-1.26.3 pillow-10.2.0 requests-2.28.1 sympy-1.12 torch-2.2.0+cu118 torchvision-0.17.0+cu118 typing-extensions-4.8.0 urllib3-1.26.13
[notice] A new release of pip available: 22.2.1 -> 24.0
[notice] To update, run: D:\GitResource\stable-diffusion-webui-directml\venv\Scripts\python.exe -m pip install --upgrade pip
Failed to automatically patch torch with ZLUDA. Could not find ZLUDA from PATH.
Traceback (most recent call last):
File "D:\GitResource\stable-diffusion-webui-directml\launch.py", line 48, in
https://github.com/lshqqytiger/stable-diffusion-webui-directml/discussions/385#discussioncomment-8476128 You can skip step 3, 4. About step 6, cuDNN will be automatically disabled.
#385 (comment) You can skip step 3, 4. About step 6, cuDNN will be automatically disabled.
Add HIP SDK and zluda directory to Path, this step i do not know how to do, could teach me, thanks
path has add success. and reboot PC, WEBUI ok, another error like this : 1、loading stable diffusion model: RuntimeError 2、RuntimeError: "LayerNormKernelImpl" not implemented for 'Half' 3、AttributeError: 'NoneType' object has no attribute 'is_sdxl'
is it matter with my AMD6500XT?
PS D:\GitResource\stable-diffusion-webui-directml> .\webui-user.bat
venv "D:\GitResource\stable-diffusion-webui-directml\venv\Scripts\Python.exe"
WARNING: ZLUDA works best with SD.Next. Please consider migrating to SD.Next.
fatal: No names found, cannot describe anything.
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: 1.8.0-RC
Commit hash: 25a3b6cbeea8a07afd5e4594afc2f1c79f41ac1a
Installing clip
Installing open_clip
Installing requirements
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning: pytorch_lightning.utilities.distributed.rank_zero_only
has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from pytorch_lightning.utilities
instead.
rank_zero_deprecation(
Launching Web UI with arguments: --use-zluda --opt-sub-quad-attention --lowvram --disable-nan-check
ONNX: selected=DmlExecutionProvider, available=['AzureExecutionProvider', 'CPUExecutionProvider']
ZLUDA device failed to pass basic operation test: index=None, device_name=AMD Radeon(TM) RX 6500 XT [ZLUDA]
Loading weights [bb32ad727a] from D:\GitResource\stable-diffusion-webui-directml\models\Stable-diffusion\darkSushi25D25D_v40.safetensors
Running on local URL: http://127.0.0.1:7860
To create a public link, set share=True
in launch()
.
Creating model from config: D:\GitResource\stable-diffusion-webui-directml\configs\v1-inference.yaml
Startup time: 73.2s (prepare environment: 74.1s, initialize shared: 3.6s, load scripts: 0.3s, create ui: 0.3s, gradio launch: 0.3s).
Couldn't find VAE named None; using None instead
Applying attention optimization: InvokeAI... done.
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "D:\Program Files\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "D:\Program Files\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "D:\Program Files\Python\Python310\lib\threading.py", line 953, in run
self._target(*self._args, self._kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\modules\initialize.py", line 148, in load_model
shared.sd_model # noqa: B018
File "D:\GitResource\stable-diffusion-webui-directml\modules\shared_items.py", line 148, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "D:\GitResource\stable-diffusion-webui-directml\modules\sd_models.py", line 627, in get_sd_model
load_model()
File "D:\GitResource\stable-diffusion-webui-directml\modules\sd_models.py", line 789, in load_model
sd_model.cond_stage_model_empty_prompt = get_empty_cond(sd_model)
File "D:\GitResource\stable-diffusion-webui-directml\modules\sd_models.py", line 665, in get_empty_cond
return sd_model.cond_stage_model([""])
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, *kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\modules\sd_hijack_clip.py", line 234, in forward
z = self.process_tokens(tokens, multipliers)
File "D:\GitResource\stable-diffusion-webui-directml\modules\sd_hijack_clip.py", line 276, in process_tokens
z = self.encode_with_transformers(tokens)
File "D:\GitResource\stable-diffusion-webui-directml\modules\sd_hijack_clip.py", line 331, in encode_with_transformers
outputs = self.wrapped.transformer(input_ids=tokens, output_hidden_states=-opts.CLIP_stop_at_last_layers)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1561, in _call_impl
result = forward_call(*args, *kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\transformers\models\clip\modeling_clip.py", line 822, in forward
return self.text_model(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(*args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\transformers\models\clip\modeling_clip.py", line 740, in forward
encoder_outputs = self.encoder(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, *kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\transformers\models\clip\modeling_clip.py", line 654, in forward
layer_outputs = encoder_layer(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(*args, *kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\transformers\models\clip\modeling_clip.py", line 382, in forward
hidden_states = self.layer_norm1(hidden_states)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(*args, **kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\extensions-builtin\Lora\networks.py", line 545, in network_LayerNorm_forward
return originals.LayerNorm_forward(self, input)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\normalization.py", line 201, in forward
return F.layer_norm(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\functional.py", line 2546, in layer_norm
return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
RuntimeError: "LayerNormKernelImpl" not implemented for 'Half'
Stable diffusion model failed to load
Traceback (most recent call last):
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\blocks.py", line 1431, in process_api
result = await self.call_function(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, args)
File "D:\GitResource\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, kwargs)
File "D:\GitResource\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 703, in pages_html
create_html()
File "D:\GitResource\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 699, in create_html
ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
File "D:\GitResource\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 699, in
RX 6500 XT is gfx1034, which is officially not supported by HIP SDK. But you can build required files yourself. And there are some people who already did it: https://github.com/brknsoul/ROCmLibs but it does not contain gfx1034 one. Related: https://www.reddit.com/r/StableDiffusion/comments/1asewch/installing_zluda_for_amd_gpus_in_windows_ie_use/
RX 6500 XT is gfx1034, which is officially not supported by HIP SDK. But you can build required files yourself. And there are some people who already did it: https://github.com/brknsoul/ROCmLibs but it does not contain gfx1034 one. Related: https://www.reddit.com/r/StableDiffusion/comments/1asewch/installing_zluda_for_amd_gpus_in_windows_ie_use/
can you help me to download Tensile-fix-fallback-arch-build.patch, i do not have right to get it
Try logging in or creating an account
the message says.
Previously running alright. Now having the same problem as LYC878484
Checklist
What happened?
change set COMMANDLINE_ARGS=--use-directml --no-half --precision full --skip-torch-cuda-test --opt-sub-quad-attention --lowvram --disable-nan-check to set COMMANDLINE_ARGS=--use-zluda --no-half --precision full --skip-torch-cuda-test --opt-sub-quad-attention --lowvram --disable-nan-check,
Steps to reproduce the problem
1、change COMMANDLINE_ARGS 2、process "webui-user.bat", and set parameters on WEBUI 3、generate picture,error: RuntimeError: "log_vml_cpu" not implemented for 'Half' Time taken: 0.8 sec.
What should have happened?
use_zluda and use_directml can use on same environment,only need to change COMMANDLINE_ARGS
What browsers do you use to access the UI ?
Google Chrome
Sysinfo
sysinfo-2024-03-17-11-50.json
Console logs
Additional information
my environment is windows10 + inter 12400F CPU + AMD6500XT