Open JohnieXu opened 7 months ago
+1
try adding --reinstall-torch to the commandline ARGS of webui. That helped me. Remember to remove it after reinstallation or it will reinstall every startup.
try adding --reinstall-torch to the commandline ARGS of webui. That helped me. Remember to remove it after reinstallation or it will reinstall every startup.
Thank u, but it doesn't work for me
################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye), Fedora 34+ and openSUSE Leap 15.4 or newer.
################################################################
################################################################
Running on xuzhiyong user
################################################################
################################################################
Repo already cloned, using it as install directory
################################################################
################################################################
Create and activate python venv
################################################################
################################################################
Launching launch.py...
################################################################
Python 3.10.6 (main, Mar 9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Installing torch and torchvision
Looking in indexes: https://pypi.tuna.tsinghua.edu.cn/simple
Requirement already satisfied: torch==2.1.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (2.1.0)
Requirement already satisfied: torchvision==0.16.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (0.16.0)
Requirement already satisfied: typing-extensions in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (4.10.0)
Requirement already satisfied: sympy in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (1.12)
Requirement already satisfied: jinja2 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.1.3)
Requirement already satisfied: filelock in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.13.1)
Requirement already satisfied: fsspec in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (2024.2.0)
Requirement already satisfied: networkx in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.2.1)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (9.5.0)
Requirement already satisfied: numpy in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (1.26.2)
Requirement already satisfied: requests in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (2.31.0)
Requirement already satisfied: MarkupSafe>=2.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from jinja2->torch==2.1.0) (2.1.5)
Requirement already satisfied: urllib3<3,>=1.21.1 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (2.2.1)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (3.3.2)
Requirement already satisfied: idna<4,>=2.5 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (3.6)
Requirement already satisfied: certifi>=2017.4.17 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (2024.2.2)
Requirement already satisfied: mpmath>=0.19 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from sympy->torch==2.1.0) (1.3.0)
[notice] A new release of pip available: 22.2.1 -> 24.0
[notice] To update, run: pip install --upgrade pip
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate --reinstall-torch
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.
Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [d124fc18f0] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/AOM3_orangemixs.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.
Startup time: 19.0s (prepare environment: 3.5s, import torch: 7.2s, import gradio: 1.9s, setup paths: 2.2s, initialize shared: 0.2s, other imports: 1.1s, load scripts: 0.9s, reload hypernetworks: 0.2s, create ui: 0.9s, gradio launch: 0.7s).
Applying attention optimization: InvokeAI... done.
loading stable diffusion model: AssertionError
Traceback (most recent call last):
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 973, in _bootstrap
self._bootstrap_inner()
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
self.run()
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
self._target(*self._args, **self._kwargs)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 148, in load_model
shared.sd_model # noqa: B018
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/shared_items.py", line 133, in sd_model
return modules.sd_models.model_data.get_sd_model()
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 621, in get_sd_model
load_model()
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 771, in load_model
with devices.autocast(), torch.no_grad():
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
if has_xpu() or has_mps() or cuda_no_autocast():
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
device_id = get_cuda_device_id()
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
) or torch.cuda.current_device()
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
_lazy_init()
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
raise AssertionError("Torch not compiled with CUDA enabled")
AssertionError: Torch not compiled with CUDA enabled
Stable diffusion model failed to load
Exception in thread Thread-2 (load_model):
Traceback (most recent call last):
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
self.run()
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
self._target(*self._args, **self._kwargs)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 153, in load_model
devices.first_time_calculation()
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 267, in first_time_calculation
linear(x)
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/networks.py", line 500, in network_Linear_forward
return originals.Linear_forward(self, input)
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 114, in forward
return F.linear(input, self.weight, self.bias)
RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'
Traceback (most recent call last):
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
result = await self.call_function(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
result = context.run(func, *args)
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
response = f(*args, **kwargs)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 703, in pages_html
create_html()
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in create_html
ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in <listcomp>
ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in create_html
self.items = {x["name"]: x for x in items_list}
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in <dictcomp>
self.items = {x["name"]: x for x in items_list}
File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 82, in list_items
item = self.create_item(name, index)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 69, in create_item
elif shared.sd_model.is_sdxl and sd_version != network.SdVersion.SDXL:
AttributeError: 'NoneType' object has no attribute 'is_sdxl'
Traceback (most recent call last):
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
output = await app.get_blocks().process_api(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
result = await self.call_function(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
prediction = await anyio.to_thread.run_sync(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
result = context.run(func, *args)
File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
response = f(*args, **kwargs)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 703, in pages_html
create_html()
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in create_html
ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in <listcomp>
ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in create_html
self.items = {x["name"]: x for x in items_list}
File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in <dictcomp>
self.items = {x["name"]: x for x in items_list}
File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 82, in list_items
item = self.create_item(name, index)
File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 69, in create_item
elif shared.sd_model.is_sdxl and sd_version != network.SdVersion.SDXL:
AttributeError: 'NoneType' object has no attribute 'is_sdxl'
hello, is this problem solved? I ran into the same problem now.
by the way, how to print the "sysinfo" in your first comment please?
hello, is this problem solved? I ran into the same problem now.
by the way, how to print the "sysinfo" in your first comment please?
not solved!
You can fetch the "sysinfo" with these steps:
hello, is this problem solved? I ran into the same problem now.
by the way, how to print the "sysinfo" in your first comment please?
After running the web UI, open the settings tab page, then click on the "Other" option on the left side and select "Sysinfo". You should then be able to download the system information from the page.
Checklist
What happened?
when i use txt2img, an error happened:
AssertionError: Torch not compiled with CUDA enable
Steps to reproduce the problem
What should have happened?
no error happened
What browsers do you use to access the UI ?
Google Chrome
Sysinfo
{ "Platform": "macOS-13.6.3-x86_64-i386-64bit", "Python": "3.10.6", "Version": "v1.8.0", "Commit": "bef51aed032c0aaa5cfd80445bc4cf0d85b408b5", "Script path": "/Volumes/RTL9210/SD/stable-diffusion-webui", "Data path": "/Volumes/RTL9210/SD/stable-diffusion-webui", "Extensions dir": "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions", "Checksum": "416a4ccbaf45286a314038b0a5e9334f25882bc163eedaff19b4b210c8053194", "Commandline": [ "launch.py", "--skip-torch-cuda-test", "--upcast-sampling", "--no-half-vae", "--use-cpu", "interrogate" ], "Torch env info": { "torch_version": "2.1.0", "is_debug_build": "False", "cuda_compiled_version": null, "gcc_version": null, "clang_version": "15.0.0 (clang-1500.0.40.1)", "cmake_version": "version 3.28.1", "os": "macOS 13.6.3 (x86_64)", "libc_version": "N/A", "python_version": "3.10.6 (main, Mar 9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)] (64-bit runtime)", "python_platform": "macOS-13.6.3-x86_64-i386-64bit", "is_cuda_available": "False", "cuda_runtime_version": null, "cuda_module_loading": "N/A", "nvidia_driver_version": null, "nvidia_gpu_models": null, "cudnn_version": null, "pip_version": "pip3", "pip_packages": [ "numpy==1.26.2", "open-clip-torch==2.20.0", "pytorch-lightning==1.9.4", "torch==2.1.0", "torchdiffeq==0.2.3", "torchmetrics==1.3.1", "torchsde==0.2.6", "torchvision==0.16.0" ], "conda_packages": null, "hip_compiled_version": "N/A", "hip_runtime_version": "N/A", "miopen_runtime_version": "N/A", "caching_allocator_config": "", "is_xnnpack_available": "True", "cpu_info": "Intel(R) Core(TM) i5-7360U CPU @ 2.30GHz" }, "Exceptions": [ { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 57, f", "res = list(func(*args, kwargs))" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 36, f", "res = func(*args, *kwargs)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py, line 110, txt2img", "processed = processing.process_images(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 785, process_images", "res = process_images_inner(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 854, process_images_inner", "with devices.autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError(\"Torch not compiled with CUDA enabled\")" ] ] }, { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 57, f", "res = list(func(args, kwargs))" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 36, f", "res = func(*args, **kwargs)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py, line 110, txt2img", "processed = processing.process_images(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 785, process_images", "res = process_images_inner(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 854, process_images_inner", "with devices.autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError(\"Torch not compiled with CUDA enabled\")" ] ] }, { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py, line 621, get_sd_model", "load_model()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py, line 771, load_model", "with devices.autocast(), torch.no_grad():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError(\"Torch not compiled with CUDA enabled\")" ] ] } ], "CPU": { "model": "i386", "count logical": 4, "count physical": 2 }, "RAM": { "total": "16GB", "used": "9GB", "free": "72MB", "active": "5GB", "inactive": "5GB" }, "Extensions": [], "Inactive extensions": [], "Environment": { "COMMANDLINE_ARGS": "--skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate", "GIT": "git", "GRADIO_ANALYTICS_ENABLED": "False", "TORCH_COMMAND": "pip install torch==2.1.0 torchvision==0.16.0" }, "Config": { "ldsr_steps": 100, "ldsr_cached": false, "SCUNET_tile": 256, "SCUNET_tile_overlap": 8, "SWIN_tile": 192, "SWIN_tile_overlap": 8, "SWIN_torch_compile": false, "hypertile_enable_unet": false, "hypertile_enable_unet_secondpass": false, "hypertile_max_depth_unet": 3, "hypertile_max_tile_unet": 256, "hypertile_swap_size_unet": 3, "hypertile_enable_vae": false, "hypertile_max_depth_vae": 3, "hypertile_max_tile_vae": 128, "hypertile_swap_size_vae": 3, "sd_model_checkpoint": "v1-5-pruned-emaonly.safetensors [6ce0161689]", "sd_checkpoint_hash": "6ce0161689b3853acaa03779ec93eafe75a02f4ced659bee03f50797806fa2fa" }, "Startup": { "total": 11.264039754867554, "records": { "initial startup": 0.01104283332824707, "prepare environment/checks": 7.796287536621094e-05, "prepare environment/git version info": 0.04249715805053711, "prepare environment/torch GPU test": 0.0039539337158203125, "prepare environment/clone repositores": 0.13756275177001953, "prepare environment/run extensions installers": 0.00048422813415527344, "prepare environment": 0.2236621379852295, "launcher": 0.0011608600616455078, "import torch": 4.559276103973389, "import gradio": 1.2403457164764404, "setup paths": 0.965001106262207, "import ldm": 0.003961086273193359, "import sgm": 5.0067901611328125e-06, "initialize shared": 0.14409089088439941, "other imports": 0.9297831058502197, "opts onchange": 0.0005729198455810547, "setup SD model": 9.393692016601562e-05, "setup codeformer": 0.0010669231414794922, "setup gfpgan": 0.006997108459472656, "set samplers": 4.601478576660156e-05, "list extensions": 0.0011892318725585938, "restore config state file": 1.0013580322265625e-05, "list SD models": 0.08728671073913574, "list localizations": 0.000331878662109375, "load scripts/custom_code.py": 0.0021550655364990234, "load scripts/img2imgalt.py": 0.00045371055603027344, "load scripts/loopback.py": 0.0002651214599609375, "load scripts/outpainting_mk_2.py": 0.0002989768981933594, "load scripts/poor_mans_outpainting.py": 0.00023317337036132812, "load scripts/postprocessing_caption.py": 0.00018787384033203125, "load scripts/postprocessing_codeformer.py": 0.00023293495178222656, "load scripts/postprocessing_create_flipped_copies.py": 0.00018787384033203125, "load scripts/postprocessing_focal_crop.py": 0.0006661415100097656, "load scripts/postprocessing_gfpgan.py": 0.00019788742065429688, "load scripts/postprocessing_split_oversized.py": 0.00019931793212890625, "load scripts/postprocessing_upscale.py": 0.0002818107604980469, "load scripts/processing_autosized_crop.py": 0.00021123886108398438, "load scripts/prompt_matrix.py": 0.0002276897430419922, "load scripts/prompts_from_file.py": 0.0003600120544433594, "load scripts/sd_upscale.py": 0.00022125244140625, "load scripts/xyz_grid.py": 0.002402782440185547, "load scripts/ldsr_model.py": 0.5824201107025146, "load scripts/lora_script.py": 0.1882781982421875, "load scripts/scunet_model.py": 0.030708789825439453, "load scripts/swinir_model.py": 0.03022289276123047, "load scripts/hotkey_config.py": 0.00026297569274902344, "load scripts/extra_options_section.py": 0.00027632713317871094, "load scripts/hypertile_script.py": 0.05973458290100098, "load scripts/hypertile_xyz.py": 0.00025534629821777344, "load scripts/soft_inpainting.py": 0.0003829002380371094, "load scripts/comments.py": 0.029378890991210938, "load scripts/refiner.py": 0.0002911090850830078, "load scripts/seed.py": 0.00023698806762695312, "load scripts": 0.9312648773193359, "load upscalers": 0.0019369125366210938, "refresh VAE": 0.0017819404602050781, "refresh textual inversion templates": 7.605552673339844e-05, "scripts list_optimizers": 0.00040721893310546875, "scripts list_unets": 6.9141387939453125e-06, "reload hypernetworks": 0.21456289291381836, "initialize extra networks": 0.03221607208251953, "scripts before_ui_callback": 0.0063779354095458984, "create ui": 1.0929930210113525, "gradio launch": 0.7930488586425781, "add APIs": 0.05200815200805664, "app_started_callback/lora_script.py": 0.0007350444793701172, "app_started_callback": 0.0007450580596923828 } }, "Packages": [ "accelerate==0.21.0", "aenum==3.1.15", "aiofiles==23.2.1", "aiohttp==3.9.3", "aiosignal==1.3.1", "altair==5.2.0", "antlr4-python3-runtime==4.9.3", "anyio==3.7.1", "async-timeout==4.0.3", "attrs==23.2.0", "blendmodes==2022", "certifi==2024.2.2", "charset-normalizer==3.3.2", "clean-fid==0.1.35", "click==8.1.7", "clip==1.0", "contourpy==1.2.0", "cycler==0.12.1", "deprecation==2.1.0", "einops==0.4.1", "exceptiongroup==1.2.0", "facexlib==0.3.0", "fastapi==0.94.0", "ffmpy==0.3.2", "filelock==3.13.1", "filterpy==1.4.5", "fonttools==4.49.0", "frozenlist==1.4.1", "fsspec==2024.2.0", "ftfy==6.1.3", "gitdb==4.0.11", "gitpython==3.1.32", "gradio-client==0.5.0", "gradio==3.41.2", "h11==0.12.0", "httpcore==0.15.0", "httpx==0.24.1", "huggingface-hub==0.21.4", "idna==3.6", "imageio==2.34.0", "importlib-resources==6.1.3", "inflection==0.5.1", "jinja2==3.1.3", "jsonmerge==1.8.0", "jsonschema-specifications==2023.12.1", "jsonschema==4.21.1", "kiwisolver==1.4.5", "kornia==0.6.7", "lark==1.1.2", "lazy-loader==0.3", "lightning-utilities==0.10.1", "llvmlite==0.42.0", "markupsafe==2.1.5", "matplotlib==3.8.3", "mpmath==1.3.0", "multidict==6.0.5", "networkx==3.2.1", "numba==0.59.0", "numpy==1.26.2", "omegaconf==2.2.3", "open-clip-torch==2.20.0", "opencv-python==4.9.0.80", "orjson==3.9.15", "packaging==23.2", "pandas==2.2.1", "piexif==1.1.3", "pillow==9.5.0", "pip==22.2.1", "protobuf==3.20.0", "psutil==5.9.5", "pydantic==1.10.14", "pydub==0.25.1", "pyparsing==3.1.2", "python-dateutil==2.9.0.post0", "python-multipart==0.0.9", "pytorch-lightning==1.9.4", "pytz==2024.1", "pywavelets==1.5.0", "pyyaml==6.0.1", "referencing==0.33.0", "regex==2023.12.25", "requests==2.31.0", "resize-right==0.0.2", "rpds-py==0.18.0", "safetensors==0.4.2", "scikit-image==0.21.0", "scipy==1.12.0", "semantic-version==2.10.0", "sentencepiece==0.2.0", "setuptools==63.2.0", "six==1.16.0", "smmap==5.0.1", "sniffio==1.3.1", "spandrel==0.1.6", "starlette==0.26.1", "sympy==1.12", "tifffile==2024.2.12", "timm==0.9.16", "tokenizers==0.13.3", "tomesd==0.1.3", "toolz==0.12.1", "torch==2.1.0", "torchdiffeq==0.2.3", "torchmetrics==1.3.1", "torchsde==0.2.6", "torchvision==0.16.0", "tqdm==4.66.2", "trampoline==0.1.2", "transformers==4.30.2", "typing-extensions==4.10.0", "tzdata==2024.1", "urllib3==2.2.1", "uvicorn==0.27.1", "wcwidth==0.2.13", "websockets==11.0.3", "yarl==1.9.4" ] }
Console logs
Additional information
No response