Closed mr-september closed 8 months ago
I'm having this issue too. It seems like it stopped working as soon as I added --use-directml to command args. It worked perfectly before, but I couldn't generate anything without directml, so I had to enable it.
Unless you are using GCN cards (~Radeon VII), ZLUDA is a better option than DirectML. If you are, revert to webui 1.7. #407
Ok, I managed to install 1.7 with the git checkout. Now I'm having trouble running it, --direct-ml won't work and I'm not sure what to do. I tried --use-zluda but it didn't work either it gives me:
Please copy full error log in the terminal. The error message in webui is less informative.
If you tried DirectML before, remove venv
folder to make sure that torch is reinstalled and patched with dll files of ZLUDA.
If you tried DirectML before, remove venv folder to make sure that torch is reinstalled and patched with dll files of ZLUDA.
is there a special command I have to enter for enabling Zluda?
I deleted the venv folder like you said and ran webui bat again and it installed what it did before and gave me this code:
`Creating venv in directory C:\AI\stable-diffusion-webui-directml\venv using python "C:\Users\rohle\AppData\Local\Programs\Python\Python310\python.exe" venv "C:\AI\stable-diffusion-webui-directml\venv\Scripts\Python.exe" fatal: No names found, cannot describe anything. Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: 1.7.0 Commit hash: 601f7e3704707d09ca88241e663a763a2493b11a Installing torch and torchvision Looking in indexes: https://pypi.org/simple, https://download.pytorch.org/whl/cu121 Collecting torch==2.2.0 Using cached https://download.pytorch.org/whl/cu121/torch-2.2.0%2Bcu121-cp310-cp310-win_amd64.whl (2454.8 MB) Collecting torchvision==0.17.0 Using cached https://download.pytorch.org/whl/cu121/torchvision-0.17.0%2Bcu121-cp310-cp310-win_amd64.whl (5.7 MB) Collecting filelock Using cached filelock-3.13.3-py3-none-any.whl (11 kB) Collecting jinja2 Using cached Jinja2-3.1.3-py3-none-any.whl (133 kB) Collecting typing-extensions>=4.8.0 Using cached typing_extensions-4.10.0-py3-none-any.whl (33 kB) Collecting sympy Using cached https://download.pytorch.org/whl/sympy-1.12-py3-none-any.whl (5.7 MB) Collecting fsspec Using cached fsspec-2024.3.1-py3-none-any.whl (171 kB) Collecting networkx Using cached https://download.pytorch.org/whl/networkx-3.2.1-py3-none-any.whl (1.6 MB) Collecting pillow!=8.3.*,>=5.3.0 Using cached https://download.pytorch.org/whl/pillow-10.2.0-cp310-cp310-win_amd64.whl (2.6 MB) Collecting numpy Using cached numpy-1.26.4-cp310-cp310-win_amd64.whl (15.8 MB) Collecting requests Using cached requests-2.31.0-py3-none-any.whl (62 kB) Collecting MarkupSafe>=2.0 Using cached MarkupSafe-2.1.5-cp310-cp310-win_amd64.whl (17 kB) Collecting idna<4,>=2.5 Using cached idna-3.6-py3-none-any.whl (61 kB) Collecting urllib3<3,>=1.21.1 Using cached urllib3-2.2.1-py3-none-any.whl (121 kB) Collecting charset-normalizer<4,>=2 Using cached charset_normalizer-3.3.2-cp310-cp310-win_amd64.whl (100 kB) Collecting certifi>=2017.4.17 Using cached certifi-2024.2.2-py3-none-any.whl (163 kB) Collecting mpmath>=0.19 Using cached https://download.pytorch.org/whl/mpmath-1.3.0-py3-none-any.whl (536 kB) Installing collected packages: mpmath, urllib3, typing-extensions, sympy, pillow, numpy, networkx, MarkupSafe, idna, fsspec, filelock, charset-normalizer, certifi, requests, jinja2, torch, torchvision Successfully installed MarkupSafe-2.1.5 certifi-2024.2.2 charset-normalizer-3.3.2 filelock-3.13.3 fsspec-2024.3.1 idna-3.6 jinja2-3.1.3 mpmath-1.3.0 networkx-3.2.1 numpy-1.26.4 pillow-10.2.0 requests-2.31.0 sympy-1.12 torch-2.2.0+cu121 torchvision-0.17.0+cu121 typing-extensions-4.10.0 urllib3-2.2.1
[notice] A new release of pip available: 22.2.1 -> 24.0
[notice] To update, run: C:\AI\stable-diffusion-webui-directml\venv\Scripts\python.exe -m pip install --upgrade pip
Installing clip
Installing open_clip
Installing requirements for CodeFormer
Installing requirements
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning: pytorch_lightning.utilities.distributed.rank_zero_only
has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from pytorch_lightning.utilities
instead.
rank_zero_deprecation(
Launching Web UI with arguments:
Style database not found: C:\AI\stable-diffusion-webui-directml\styles.csv
ONNX: selected=CUDAExecutionProvider, available=['AzureExecutionProvider', 'CPUExecutionProvider']
Loading weights [fe4efff1e1] from C:\AI\stable-diffusion-webui-directml\models\Stable-diffusion\sd-v1-4.ckpt
Running on local URL: http://127.0.0.1:7860
To create a public link, set share=True
in launch()
.
Startup time: 230.5s (prepare environment: 237.2s, initialize shared: 1.8s, load scripts: 0.5s, create ui: 0.5s, gradio launch: 0.2s).
Creating model from config: C:\AI\stable-diffusion-webui-directml\configs\v1-inference.yaml
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 953, in run
self._target(*self._args, self._kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\initialize.py", line 147, in load_model
shared.sd_model # noqa: B018
File "C:\AI\stable-diffusion-webui-directml\modules\shared_items.py", line 143, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 537, in get_sd_model
load_model()
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 675, in load_model
load_model_weights(sd_model, checkpoint_info, state_dict, timer)
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 381, in load_model_weights
model.load_state_dict(state_dict, strict=False)
File "C:\AI\stable-diffusion-webui-directml\modules\sd_disable_initialization.py", line 223, in TORCH_USE_CUDA_DSA
to enable device-side assertions.
Stable diffusion model failed to load
Applying attention optimization: Doggettx... done.
Loading weights [fe4efff1e1] from C:\AI\stable-diffusion-webui-directml\models\Stable-diffusion\sd-v1-4.ckpt
Exception in thread Thread-26 (load_model):
Traceback (most recent call last):
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 953, in run
self._target(*self._args, self._kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\initialize.py", line 153, in load_model
devices.first_time_calculation()
File "C:\AI\stable-diffusion-webui-directml\modules\devices.py", line 178, in first_time_calculation
linear(x)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl
return self._call_impl(*args, *kwargs)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl
return forward_call(args, kwargs)
File "C:\AI\stable-diffusion-webui-directml\extensions-builtin\Lora\networks.py", line 486, in network_Linear_forward
return originals.Linear_forward(self, input)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\linear.py", line 116, in forward
return F.linear(input, self.weight, self.bias)
RuntimeError: CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling cublasGemmEx( handle, opa, opb, m, n, k, &falpha, a, CUDA_R_16F, lda, b, CUDA_R_16F, ldb, &fbeta, c, CUDA_R_16F, ldc, CUDA_R_32F, CUBLAS_GEMM_DEFAULT_TENSOR_OP)
Creating model from config: C:\AI\stable-diffusion-webui-directml\configs\v1-inference.yaml
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, args)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 419, in pages_html
return refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 425, in refresh
pg.refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks_textual_inversion.py", line 15, in refresh
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 224, in load_textual_inversion_embeddings
self.expected_shape = self.get_expected_shape()
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 156, in get_expected_shape
vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
File "C:\AI\stable-diffusion-webui-directml\modules\shared_items.py", line 143, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 537, in get_sd_model
load_model()
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 675, in load_model
load_model_weights(sd_model, checkpoint_info, state_dict, timer)
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 381, in load_model_weights
model.load_state_dict(state_dict, strict=False)
File "C:\AI\stable-diffusion-webui-directml\modules\sd_disable_initialization.py", line 223, in TORCH_USE_CUDA_DSA
to enable device-side assertions.
Stable diffusion model failed to load
Loading weights [fe4efff1e1] from C:\AI\stable-diffusion-webui-directml\models\Stable-diffusion\sd-v1-4.ckpt
Traceback (most recent call last):
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\blocks.py", line 1431, in process_api
result = await self.call_function(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, args)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 419, in pages_html
return refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 425, in refresh
pg.refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks_textual_inversion.py", line 15, in refresh
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 224, in load_textual_inversion_embeddings
self.expected_shape = self.get_expected_shape()
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 156, in get_expected_shape
vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
Creating model from config: C:\AI\stable-diffusion-webui-directml\configs\v1-inference.yaml
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, args)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 419, in pages_html
return refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 425, in refresh
pg.refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks_textual_inversion.py", line 15, in refresh
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 224, in load_textual_inversion_embeddings
self.expected_shape = self.get_expected_shape()
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 156, in get_expected_shape
vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
File "C:\AI\stable-diffusion-webui-directml\modules\shared_items.py", line 143, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 537, in get_sd_model
load_model()
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 675, in load_model
load_model_weights(sd_model, checkpoint_info, state_dict, timer)
File "C:\AI\stable-diffusion-webui-directml\modules\sd_models.py", line 381, in load_model_weights
model.load_state_dict(state_dict, strict=False)
File "C:\AI\stable-diffusion-webui-directml\modules\sd_disable_initialization.py", line 223, in TORCH_USE_CUDA_DSA
to enable device-side assertions.
Stable diffusion model failed to load
Loading weights [fe4efff1e1] from C:\AI\stable-diffusion-webui-directml\models\Stable-diffusion\sd-v1-4.ckpt
Traceback (most recent call last):
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\blocks.py", line 1431, in process_api
result = await self.call_function(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, args)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 419, in pages_html
return refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks.py", line 425, in refresh
pg.refresh()
File "C:\AI\stable-diffusion-webui-directml\modules\ui_extra_networks_textual_inversion.py", line 15, in refresh
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings(force_reload=True)
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 224, in load_textual_inversion_embeddings
self.expected_shape = self.get_expected_shape()
File "C:\AI\stable-diffusion-webui-directml\modules\textual_inversion\textual_inversion.py", line 156, in get_expected_shape
vec = shared.sd_model.cond_stage_model.encode_embedding_init_text(",", 1)
AttributeError: 'NoneType' object has no attribute 'cond_stage_model'
Creating model from config: C:\AI\stable-diffusion-webui-directml\configs\v1-inference.yaml
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "C:\Users\rohle\AppData\Local\Programs\Python\Python310\lib\threading.py", line 1016, in _bootstrap_inner
self.run()
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\anyio_backends_asyncio.py", line 807, in run
result = context.run(func, args)
File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\gradio\utils.py", line 707, in wrapper
response = f(args, kwargs)
File "C:\AI\stable-diffusion-webui-directml\modules\ui.py", line 1816, in TORCH_USE_CUDA_DSA
to enable device-side assertions.
Stable diffusion model failed to load`
Got an amd card the 7800xt so it should be able to do zluda. I installed it earlier today and added it to path, so i don't think it's giving me errors related to that.
You should be on the latest commit. Do not checkout older commits if you are trying ZLUDA. Pull the latest commits and try again. There was an issue with ZLUDA installation. If you don't want to remove the whole venv folder,
.\venv\Scripts\activate
pip uninstall torch torchvision -y
then launch webui.
Well, I went back to 1.7 because gfpen and codeformer isn't working on 1.8. I don't wanna go back to 1.8 if this Is going to be the case. You think using zluda will fix it?
The reason that gfpgan and codeformers are broken is because DirectML does not support storage access. ZLUDA supports storage access, so there should not be a problem. However, ZLUDA does not have hiprtc that is required by codeformers. Therefore, gfpgan should work well, and codeformers will use cpu regardless of 1.7 or 1.8. In short, DirectML + webui 1.8: both don't work. DirectML + webui 1.7: both work. ZLUDA: best performance. but codeformers uses cpu.
I'm in the middle of installing 1.8 with Zluda. First run takes some time, but I'm hopeful.
Ok, so I got codeformer working! thank you! The only problem I have now is that it says no GFPGAN model found when i select it:
WARNING:modules.face_restoration_utils:Unable to load face-restoration model Traceback (most recent call last): File "C:\AI\stable-diffusion-webui-directml\modules\face_restoration_utils.py", line 150, in restore_with_helper self.net = self.load_net() File "C:\AI\stable-diffusion-webui-directml\modules\gfpgan_model.py", line 46, in load_net raise ValueError("No GFPGAN model found") ValueError: No GFPGAN model found
I downloaded the model manually and placed it in models/GFPGAN but then it gave me a different error:
Failed face-restoration inference Traceback (most recent call last): File "C:\AI\stable-diffusion-webui-directml\modules\face_restoration_utils.py", line 86, in restore_with_face_helper cropped_face_t = restore_face(cropped_face_t) File "C:\AI\stable-diffusion-webui-directml\modules\gfpgan_model.py", line 51, in restore_face return self.net(cropped_face_t, return_rgb=False)[0] File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl return self._call_impl(args, kwargs) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl return forward_call(*args, kwargs) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\spandrel\architectures\GFPGAN\arch\gfpganv1_clean_arch.py", line 319, in forward feat = F.leakyrelu(self.conv_body_first(x), negative_slope=0.2) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl return forward_call(args, kwargs) File "C:\AI\stable-diffusion-webui-directml\extensions-builtin\Lora\networks.py", line 515, in network_Conv2d_forward return originals.Conv2d_forward(self, input) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\conv.py", line 460, in forward return self._conv_forward(input, self.weight, self.bias) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\conv.py", line 456, in _conv_forward return F.conv2d(input, weight, bias, self.stride, RuntimeError: Input type (torch.FloatTensor) and weight type (torch.cuda.FloatTensor) should be the same or input should be a MKLDNN tensor and weight is a dense tensor
Any ideas?
Unless you are using GCN cards (~Radeon VII), ZLUDA is a better option than DirectML. If you are, revert to webui 1.7. #407
According to this page, my GPU (RX 5700) is not supported. Strangely it is a ROCm compatibility list, not a ZLUDA, but it's directly linked from this ZLUDA page.
So I should roll back to 1.7? Or have people reported unofficial compatibility with 5700?
ZLUDA depends on ROCm, so we can use torch although we don't have torch built with ROCm on Windows. You have to build required libraries with gfx1010, but there's user-built libraries: https://github.com/brknsoul/ROCmLibs
So you are recommending I download this zip and try ZLUDA? Are there any further instructions? Where do I put the files in this zip?
So you are recommending I download this zip and try ZLUDA? Are there any further instructions? Where do I put the files in this zip?
Follow this video for the zluda guide once you build the libraries. I don't know how to build libraries.
Considering my options, in the long term I want to switch to Zluda with SD Next. It is so far impossible for me to install due to caffe2_nvrtc.dll
errors.
In the short term, to retain compatibility with this A1111 fork and updated versions, I have just added --use-cpu gfpgan codeformers
. There's no real difference in speed anyway.
Ok, so I got codeformer working! thank you! The only problem I have now is that it says no GFPGAN model found when i select it:
WARNING:modules.face_restoration_utils:Unable to load face-restoration model Traceback (most recent call last): File "C:\AI\stable-diffusion-webui-directml\modules\face_restoration_utils.py", line 150, in restore_with_helper self.net = self.load_net() File "C:\AI\stable-diffusion-webui-directml\modules\gfpgan_model.py", line 46, in load_net raise ValueError("No GFPGAN model found") ValueError: No GFPGAN model found
I downloaded the model manually and placed it in models/GFPGAN but then it gave me a different error:
Failed face-restoration inference Traceback (most recent call last): File "C:\AI\stable-diffusion-webui-directml\modules\face_restoration_utils.py", line 86, in restore_with_face_helper cropped_face_t = restore_face(cropped_face_t) File "C:\AI\stable-diffusion-webui-directml\modules\gfpgan_model.py", line 51, in restore_face return self.net(cropped_face_t, return_rgb=False)[0] File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl return self._call_impl(args, kwargs) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl return forward_call(*args, kwargs) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\spandrel\architectures\GFPGAN\arch\gfpganv1_clean_arch.py", line 319, in forward feat = F.leakyrelu(self.conv_body_first(x), negative_slope=0.2) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl return forward_call(args, kwargs) File "C:\AI\stable-diffusion-webui-directml\extensions-builtin\Lora\networks.py", line 515, in network_Conv2d_forward return originals.Conv2d_forward(self, input) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\conv.py", line 460, in forward return self._conv_forward(input, self.weight, self.bias) File "C:\AI\stable-diffusion-webui-directml\venv\lib\site-packages\torch\nn\modules\conv.py", line 456, in _conv_forward return F.conv2d(input, weight, bias, self.stride, RuntimeError: Input type (torch.FloatTensor) and weight type (torch.cuda.FloatTensor) should be the same or input should be a MKLDNN tensor and weight is a dense tensor
Any ideas?
Have the same error here with Extra tab > GFPGAN Failed face-restoration inference Traceback (most recent call last): File "F:\sd\sd_zluda\modules\face_restoration_utils.py", line 86, in restore_with_face_helper cropped_face_t = restore_face(cropped_face_t) File "F:\sd\sd_zluda\modules\gfpgan_model.py", line 51, in restore_face return self.net(cropped_face_t, return_rgb=False)[0] File "F:\sd\sd_zluda\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl return self._call_impl(args, kwargs) File "F:\sd\sd_zluda\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl return forward_call(*args, kwargs) File "F:\sd\sd_zluda\venv\lib\site-packages\spandrel\architectures\GFPGAN\arch\gfpganv1_clean_arch.py", line 319, in forward feat = F.leakyrelu(self.conv_body_first(x), negative_slope=0.2) File "F:\sd\sd_zluda\venv\lib\site-packages\torch\nn\modules\module.py", line 1511, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "F:\sd\sd_zluda\venv\lib\site-packages\torch\nn\modules\module.py", line 1520, in _call_impl return forward_call(args, kwargs) File "F:\sd\sd_zluda\extensions-builtin\Lora\networks.py", line 518, in network_Conv2d_forward return originals.Conv2d_forward(self, input) File "F:\sd\sd_zluda\venv\lib\site-packages\torch\nn\modules\conv.py", line 460, in forward return self._conv_forward(input, self.weight, self.bias) File "F:\sd\sd_zluda\venv\lib\site-packages\torch\nn\modules\conv.py", line 456, in _conv_forward return F.conv2d(input, weight, bias, self.stride, RuntimeError: Input type (torch.FloatTensor) and weight type (torch.cuda.FloatTensor) should be the same or input should be a MKLDNN tensor and weight is a dense tensor
Using rx 6600xt Version: v1.9.3-amd-4-g89e60b8e Commit hash: 89e60b8e96d1e2697af3d86e0322ff5ec10db51d ZLUDA
Checklist
What happened?
Extras tab, GFPGan and Codeformers have stopped working. I have been using them fine for months, until recently.
The output is the same as the input. There is an error message in the console, which doesn't seem to fully halt SD.
Steps to reproduce the problem
Use either/both of GFPGan and Codeformers in the Extras tab.
What should have happened?
Output with clear HD faces
What browsers do you use to access the UI ?
Google Chrome
Sysinfo
Console logs
Additional information
Have tried deleting venv, generating new venv, updating everything in requirements.txt did not work. Tried a fresh parallel install, also did not work.