AUTOMATIC1111 / stable-diffusion-webui

Stable Diffusion web UI
GNU Affero General Public License v3.0
140.74k stars 26.63k forks source link

[Bug]: AssertionError: Torch not compiled with CUDA enable #15192

Open JohnieXu opened 7 months ago

JohnieXu commented 7 months ago

Checklist

What happened?

when i use txt2img, an error happened: AssertionError: Torch not compiled with CUDA enable

Steps to reproduce the problem

Python 3.10.6 (main, Mar  9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.

Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [6ce0161689] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/v1-5-pruned-emaonly.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

What should have happened?

no error happened

What browsers do you use to access the UI ?

Google Chrome

Sysinfo

{ "Platform": "macOS-13.6.3-x86_64-i386-64bit", "Python": "3.10.6", "Version": "v1.8.0", "Commit": "bef51aed032c0aaa5cfd80445bc4cf0d85b408b5", "Script path": "/Volumes/RTL9210/SD/stable-diffusion-webui", "Data path": "/Volumes/RTL9210/SD/stable-diffusion-webui", "Extensions dir": "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions", "Checksum": "416a4ccbaf45286a314038b0a5e9334f25882bc163eedaff19b4b210c8053194", "Commandline": [ "launch.py", "--skip-torch-cuda-test", "--upcast-sampling", "--no-half-vae", "--use-cpu", "interrogate" ], "Torch env info": { "torch_version": "2.1.0", "is_debug_build": "False", "cuda_compiled_version": null, "gcc_version": null, "clang_version": "15.0.0 (clang-1500.0.40.1)", "cmake_version": "version 3.28.1", "os": "macOS 13.6.3 (x86_64)", "libc_version": "N/A", "python_version": "3.10.6 (main, Mar 9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)] (64-bit runtime)", "python_platform": "macOS-13.6.3-x86_64-i386-64bit", "is_cuda_available": "False", "cuda_runtime_version": null, "cuda_module_loading": "N/A", "nvidia_driver_version": null, "nvidia_gpu_models": null, "cudnn_version": null, "pip_version": "pip3", "pip_packages": [ "numpy==1.26.2", "open-clip-torch==2.20.0", "pytorch-lightning==1.9.4", "torch==2.1.0", "torchdiffeq==0.2.3", "torchmetrics==1.3.1", "torchsde==0.2.6", "torchvision==0.16.0" ], "conda_packages": null, "hip_compiled_version": "N/A", "hip_runtime_version": "N/A", "miopen_runtime_version": "N/A", "caching_allocator_config": "", "is_xnnpack_available": "True", "cpu_info": "Intel(R) Core(TM) i5-7360U CPU @ 2.30GHz" }, "Exceptions": [ { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 57, f", "res = list(func(*args, kwargs))" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 36, f", "res = func(*args, *kwargs)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py, line 110, txt2img", "processed = processing.process_images(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 785, process_images", "res = process_images_inner(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 854, process_images_inner", "with devices.autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError(\"Torch not compiled with CUDA enabled\")" ] ] }, { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 57, f", "res = list(func(args, kwargs))" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 36, f", "res = func(*args, **kwargs)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py, line 110, txt2img", "processed = processing.process_images(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 785, process_images", "res = process_images_inner(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 854, process_images_inner", "with devices.autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError(\"Torch not compiled with CUDA enabled\")" ] ] }, { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py, line 621, get_sd_model", "load_model()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py, line 771, load_model", "with devices.autocast(), torch.no_grad():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError(\"Torch not compiled with CUDA enabled\")" ] ] } ], "CPU": { "model": "i386", "count logical": 4, "count physical": 2 }, "RAM": { "total": "16GB", "used": "9GB", "free": "72MB", "active": "5GB", "inactive": "5GB" }, "Extensions": [], "Inactive extensions": [], "Environment": { "COMMANDLINE_ARGS": "--skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate", "GIT": "git", "GRADIO_ANALYTICS_ENABLED": "False", "TORCH_COMMAND": "pip install torch==2.1.0 torchvision==0.16.0" }, "Config": { "ldsr_steps": 100, "ldsr_cached": false, "SCUNET_tile": 256, "SCUNET_tile_overlap": 8, "SWIN_tile": 192, "SWIN_tile_overlap": 8, "SWIN_torch_compile": false, "hypertile_enable_unet": false, "hypertile_enable_unet_secondpass": false, "hypertile_max_depth_unet": 3, "hypertile_max_tile_unet": 256, "hypertile_swap_size_unet": 3, "hypertile_enable_vae": false, "hypertile_max_depth_vae": 3, "hypertile_max_tile_vae": 128, "hypertile_swap_size_vae": 3, "sd_model_checkpoint": "v1-5-pruned-emaonly.safetensors [6ce0161689]", "sd_checkpoint_hash": "6ce0161689b3853acaa03779ec93eafe75a02f4ced659bee03f50797806fa2fa" }, "Startup": { "total": 11.264039754867554, "records": { "initial startup": 0.01104283332824707, "prepare environment/checks": 7.796287536621094e-05, "prepare environment/git version info": 0.04249715805053711, "prepare environment/torch GPU test": 0.0039539337158203125, "prepare environment/clone repositores": 0.13756275177001953, "prepare environment/run extensions installers": 0.00048422813415527344, "prepare environment": 0.2236621379852295, "launcher": 0.0011608600616455078, "import torch": 4.559276103973389, "import gradio": 1.2403457164764404, "setup paths": 0.965001106262207, "import ldm": 0.003961086273193359, "import sgm": 5.0067901611328125e-06, "initialize shared": 0.14409089088439941, "other imports": 0.9297831058502197, "opts onchange": 0.0005729198455810547, "setup SD model": 9.393692016601562e-05, "setup codeformer": 0.0010669231414794922, "setup gfpgan": 0.006997108459472656, "set samplers": 4.601478576660156e-05, "list extensions": 0.0011892318725585938, "restore config state file": 1.0013580322265625e-05, "list SD models": 0.08728671073913574, "list localizations": 0.000331878662109375, "load scripts/custom_code.py": 0.0021550655364990234, "load scripts/img2imgalt.py": 0.00045371055603027344, "load scripts/loopback.py": 0.0002651214599609375, "load scripts/outpainting_mk_2.py": 0.0002989768981933594, "load scripts/poor_mans_outpainting.py": 0.00023317337036132812, "load scripts/postprocessing_caption.py": 0.00018787384033203125, "load scripts/postprocessing_codeformer.py": 0.00023293495178222656, "load scripts/postprocessing_create_flipped_copies.py": 0.00018787384033203125, "load scripts/postprocessing_focal_crop.py": 0.0006661415100097656, "load scripts/postprocessing_gfpgan.py": 0.00019788742065429688, "load scripts/postprocessing_split_oversized.py": 0.00019931793212890625, "load scripts/postprocessing_upscale.py": 0.0002818107604980469, "load scripts/processing_autosized_crop.py": 0.00021123886108398438, "load scripts/prompt_matrix.py": 0.0002276897430419922, "load scripts/prompts_from_file.py": 0.0003600120544433594, "load scripts/sd_upscale.py": 0.00022125244140625, "load scripts/xyz_grid.py": 0.002402782440185547, "load scripts/ldsr_model.py": 0.5824201107025146, "load scripts/lora_script.py": 0.1882781982421875, "load scripts/scunet_model.py": 0.030708789825439453, "load scripts/swinir_model.py": 0.03022289276123047, "load scripts/hotkey_config.py": 0.00026297569274902344, "load scripts/extra_options_section.py": 0.00027632713317871094, "load scripts/hypertile_script.py": 0.05973458290100098, "load scripts/hypertile_xyz.py": 0.00025534629821777344, "load scripts/soft_inpainting.py": 0.0003829002380371094, "load scripts/comments.py": 0.029378890991210938, "load scripts/refiner.py": 0.0002911090850830078, "load scripts/seed.py": 0.00023698806762695312, "load scripts": 0.9312648773193359, "load upscalers": 0.0019369125366210938, "refresh VAE": 0.0017819404602050781, "refresh textual inversion templates": 7.605552673339844e-05, "scripts list_optimizers": 0.00040721893310546875, "scripts list_unets": 6.9141387939453125e-06, "reload hypernetworks": 0.21456289291381836, "initialize extra networks": 0.03221607208251953, "scripts before_ui_callback": 0.0063779354095458984, "create ui": 1.0929930210113525, "gradio launch": 0.7930488586425781, "add APIs": 0.05200815200805664, "app_started_callback/lora_script.py": 0.0007350444793701172, "app_started_callback": 0.0007450580596923828 } }, "Packages": [ "accelerate==0.21.0", "aenum==3.1.15", "aiofiles==23.2.1", "aiohttp==3.9.3", "aiosignal==1.3.1", "altair==5.2.0", "antlr4-python3-runtime==4.9.3", "anyio==3.7.1", "async-timeout==4.0.3", "attrs==23.2.0", "blendmodes==2022", "certifi==2024.2.2", "charset-normalizer==3.3.2", "clean-fid==0.1.35", "click==8.1.7", "clip==1.0", "contourpy==1.2.0", "cycler==0.12.1", "deprecation==2.1.0", "einops==0.4.1", "exceptiongroup==1.2.0", "facexlib==0.3.0", "fastapi==0.94.0", "ffmpy==0.3.2", "filelock==3.13.1", "filterpy==1.4.5", "fonttools==4.49.0", "frozenlist==1.4.1", "fsspec==2024.2.0", "ftfy==6.1.3", "gitdb==4.0.11", "gitpython==3.1.32", "gradio-client==0.5.0", "gradio==3.41.2", "h11==0.12.0", "httpcore==0.15.0", "httpx==0.24.1", "huggingface-hub==0.21.4", "idna==3.6", "imageio==2.34.0", "importlib-resources==6.1.3", "inflection==0.5.1", "jinja2==3.1.3", "jsonmerge==1.8.0", "jsonschema-specifications==2023.12.1", "jsonschema==4.21.1", "kiwisolver==1.4.5", "kornia==0.6.7", "lark==1.1.2", "lazy-loader==0.3", "lightning-utilities==0.10.1", "llvmlite==0.42.0", "markupsafe==2.1.5", "matplotlib==3.8.3", "mpmath==1.3.0", "multidict==6.0.5", "networkx==3.2.1", "numba==0.59.0", "numpy==1.26.2", "omegaconf==2.2.3", "open-clip-torch==2.20.0", "opencv-python==4.9.0.80", "orjson==3.9.15", "packaging==23.2", "pandas==2.2.1", "piexif==1.1.3", "pillow==9.5.0", "pip==22.2.1", "protobuf==3.20.0", "psutil==5.9.5", "pydantic==1.10.14", "pydub==0.25.1", "pyparsing==3.1.2", "python-dateutil==2.9.0.post0", "python-multipart==0.0.9", "pytorch-lightning==1.9.4", "pytz==2024.1", "pywavelets==1.5.0", "pyyaml==6.0.1", "referencing==0.33.0", "regex==2023.12.25", "requests==2.31.0", "resize-right==0.0.2", "rpds-py==0.18.0", "safetensors==0.4.2", "scikit-image==0.21.0", "scipy==1.12.0", "semantic-version==2.10.0", "sentencepiece==0.2.0", "setuptools==63.2.0", "six==1.16.0", "smmap==5.0.1", "sniffio==1.3.1", "spandrel==0.1.6", "starlette==0.26.1", "sympy==1.12", "tifffile==2024.2.12", "timm==0.9.16", "tokenizers==0.13.3", "tomesd==0.1.3", "toolz==0.12.1", "torch==2.1.0", "torchdiffeq==0.2.3", "torchmetrics==1.3.1", "torchsde==0.2.6", "torchvision==0.16.0", "tqdm==4.66.2", "trampoline==0.1.2", "transformers==4.30.2", "typing-extensions==4.10.0", "tzdata==2024.1", "urllib3==2.2.1", "uvicorn==0.27.1", "wcwidth==0.2.13", "websockets==11.0.3", "yarl==1.9.4" ] }

Console logs

stable-diffusion-webui git:(master) ./webui.sh

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye), Fedora 34+ and openSUSE Leap 15.4 or newer.
################################################################

################################################################
Running on xuzhiyong user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
Create and activate python venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.6 (main, Mar  9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.

Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [6ce0161689] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/v1-5-pruned-emaonly.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 11.3s (prepare environment: 0.2s, import torch: 4.6s, import gradio: 1.2s, setup paths: 1.0s, initialize shared: 0.1s, other imports: 0.9s, load scripts: 0.9s, reload hypernetworks: 0.2s, create ui: 1.1s, gradio launch: 0.8s).
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
Applying attention optimization: InvokeAI... done.
loading stable diffusion model: AssertionError
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 148, in load_model
    shared.sd_model  # noqa: B018
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/shared_items.py", line 133, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 621, in get_sd_model
    load_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 771, in load_model
    with devices.autocast(), torch.no_grad():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
    if has_xpu() or has_mps() or cuda_no_autocast():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
    device_id = get_cuda_device_id()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
    ) or torch.cuda.current_device()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
    _lazy_init()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
    raise AssertionError("Torch not compiled with CUDA enabled")
AssertionError: Torch not compiled with CUDA enabled

Stable diffusion model failed to load
Using already loaded model v1-5-pruned-emaonly.safetensors [6ce0161689]: done in 0.0s
*** Error completing request
*** Arguments: ('task(6gssuckqcqqbe6l)', <gradio.routes.Request object at 0x15305ae30>, 'AssertionError: Torch not compiled with CUDA enabled', '', [], 20, 'DPM++ 2M Karras', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
    Traceback (most recent call last):
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 57, in f
        res = list(func(*args, **kwargs))
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 36, in f
        res = func(*args, **kwargs)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py", line 110, in txt2img
        processed = processing.process_images(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 785, in process_images
        res = process_images_inner(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 854, in process_images_inner
        with devices.autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
        if has_xpu() or has_mps() or cuda_no_autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
        device_id = get_cuda_device_id()
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
        ) or torch.cuda.current_device()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
        _lazy_init()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
        raise AssertionError("Torch not compiled with CUDA enabled")
    AssertionError: Torch not compiled with CUDA enabled

---
Exception in thread Thread-2 (load_model):
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 153, in load_model
    devices.first_time_calculation()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 267, in first_time_calculation
    linear(x)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
    return forward_call(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/networks.py", line 500, in network_Linear_forward
    return originals.Linear_forward(self, input)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 114, in forward
    return F.linear(input, self.weight, self.bias)
RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'
*** Error completing request
*** Arguments: ('task(2l51bhlon8pmei9)', <gradio.routes.Request object at 0x15305abf0>, 'AssertionError: Torch not compiled with CUDA enabled', '', [], 20, 'DPM++ 2M Karras', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
    Traceback (most recent call last):
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 57, in f
        res = list(func(*args, **kwargs))
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 36, in f
        res = func(*args, **kwargs)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py", line 110, in txt2img
        processed = processing.process_images(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 785, in process_images
        res = process_images_inner(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 854, in process_images_inner
        with devices.autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
        if has_xpu() or has_mps() or cuda_no_autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
        device_id = get_cuda_device_id()
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
        ) or torch.cuda.current_device()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
        _lazy_init()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
        raise AssertionError("Torch not compiled with CUDA enabled")
    AssertionError: Torch not compiled with CUDA enabled

---

Additional information

No response

huanglongyu commented 6 months ago

+1

Xavia1991 commented 6 months ago

try adding --reinstall-torch to the commandline ARGS of webui. That helped me. Remember to remove it after reinstallation or it will reinstall every startup.

JohnieXu commented 6 months ago

try adding --reinstall-torch to the commandline ARGS of webui. That helped me. Remember to remove it after reinstallation or it will reinstall every startup.

Thank u, but it doesn't work for me

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye), Fedora 34+ and openSUSE Leap 15.4 or newer.
################################################################

################################################################
Running on xuzhiyong user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
Create and activate python venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.6 (main, Mar  9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Installing torch and torchvision
Looking in indexes: https://pypi.tuna.tsinghua.edu.cn/simple
Requirement already satisfied: torch==2.1.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (2.1.0)
Requirement already satisfied: torchvision==0.16.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (0.16.0)
Requirement already satisfied: typing-extensions in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (4.10.0)
Requirement already satisfied: sympy in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (1.12)
Requirement already satisfied: jinja2 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.1.3)
Requirement already satisfied: filelock in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.13.1)
Requirement already satisfied: fsspec in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (2024.2.0)
Requirement already satisfied: networkx in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.2.1)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (9.5.0)
Requirement already satisfied: numpy in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (1.26.2)
Requirement already satisfied: requests in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (2.31.0)
Requirement already satisfied: MarkupSafe>=2.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from jinja2->torch==2.1.0) (2.1.5)
Requirement already satisfied: urllib3<3,>=1.21.1 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (2.2.1)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (3.3.2)
Requirement already satisfied: idna<4,>=2.5 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (3.6)
Requirement already satisfied: certifi>=2017.4.17 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (2024.2.2)
Requirement already satisfied: mpmath>=0.19 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from sympy->torch==2.1.0) (1.3.0)

[notice] A new release of pip available: 22.2.1 -> 24.0
[notice] To update, run: pip install --upgrade pip
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate --reinstall-torch
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.

Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [d124fc18f0] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/AOM3_orangemixs.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 19.0s (prepare environment: 3.5s, import torch: 7.2s, import gradio: 1.9s, setup paths: 2.2s, initialize shared: 0.2s, other imports: 1.1s, load scripts: 0.9s, reload hypernetworks: 0.2s, create ui: 0.9s, gradio launch: 0.7s).
Applying attention optimization: InvokeAI... done.
loading stable diffusion model: AssertionError
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 148, in load_model
    shared.sd_model  # noqa: B018
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/shared_items.py", line 133, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 621, in get_sd_model
    load_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 771, in load_model
    with devices.autocast(), torch.no_grad():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
    if has_xpu() or has_mps() or cuda_no_autocast():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
    device_id = get_cuda_device_id()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
    ) or torch.cuda.current_device()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
    _lazy_init()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
    raise AssertionError("Torch not compiled with CUDA enabled")
AssertionError: Torch not compiled with CUDA enabled

Stable diffusion model failed to load
Exception in thread Thread-2 (load_model):
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 153, in load_model
    devices.first_time_calculation()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 267, in first_time_calculation
    linear(x)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
    return forward_call(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/networks.py", line 500, in network_Linear_forward
    return originals.Linear_forward(self, input)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 114, in forward
    return F.linear(input, self.weight, self.bias)
RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 703, in pages_html
    create_html()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in create_html
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in <listcomp>
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in create_html
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in <dictcomp>
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 82, in list_items
    item = self.create_item(name, index)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 69, in create_item
    elif shared.sd_model.is_sdxl and sd_version != network.SdVersion.SDXL:
AttributeError: 'NoneType' object has no attribute 'is_sdxl'
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 703, in pages_html
    create_html()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in create_html
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in <listcomp>
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in create_html
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in <dictcomp>
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 82, in list_items
    item = self.create_item(name, index)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 69, in create_item
    elif shared.sd_model.is_sdxl and sd_version != network.SdVersion.SDXL:
AttributeError: 'NoneType' object has no attribute 'is_sdxl'
zxtgool commented 2 months ago

hello, is this problem solved? I ran into the same problem now.

by the way, how to print the "sysinfo" in your first comment please?

JohnieXu commented 2 months ago

hello, is this problem solved? I ran into the same problem now.

by the way, how to print the "sysinfo" in your first comment please?

not solved!

You can fetch the "sysinfo" with these steps:

  1. Go to WebUI Settings -> Sysinfo -> Download system info. If WebUI fails to launch, use --dump-sysinfo commandline argument to generate the file
  2. Upload the Sysinfo as a attached file, Do NOT paste it in as plain text.
wywanchongshan commented 2 weeks ago

hello, is this problem solved? I ran into the same problem now.

by the way, how to print the "sysinfo" in your first comment please?

After running the web UI, open the settings tab page, then click on the "Other" option on the left side and select "Sysinfo". You should then be able to download the system information from the page.