Open kairin opened 3 hours ago
so the first thing i did was tried to follow a suggestion to go to an older version of torch...
so i did.
ddd venv
Command 'nvcc' not found, but can be installed with:
sudo apt install nvidia-cuda-toolkit
Fri Oct 25 22:58:42 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.51.01 Driver Version: 565.90 CUDA Version: 12.7 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 4080 ... On | 00000000:0B:00.0 On | N/A |
| 0% 39C P2 17W / 320W | 2641MiB / 16376MiB | 7% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| 0 N/A N/A 35 G /Xwayland N/A |
| 0 N/A N/A 620148 C /python3.12 N/A |
+-----------------------------------------------------------------------------------------+
VIRTUAL_ENV=/home/ks/venv
VIRTUAL_ENV_PROMPT=(venv)
ddd venv
CODEOWNERS app comfyui.prev2.log fix_torch.py models pytest.ini tests-unit
CONTRIBUTING.md comfy comfyui_screenshot.png folder_paths.py new_updater.py requirements.txt user
LICENSE comfy_execution cuda_malloc.py input node_helpers.py script_examples utils
README.md comfy_extras custom_nodes latent_preview.py nodes.py server.py web
__pycache__ comfyui.log execution.py main.py notebooks temp
api_server comfyui.prev.log extra_model_paths.yaml.example model_filemanager output tests
on bash ks 3.12.3 master ≡ 0ms
╭─ 22:58:43 | 25 Oct, Friday | in venv comf
╰─❯ pip install torch==2.4.1 torchvision==0.19.1 torchaudio==2.4.1 --index-url https://download.pytorch.org/whl/cu124
Looking in indexes: https://download.pytorch.org/whl/cu124
Collecting torch==2.4.1
Downloading https://download.pytorch.org/whl/cu124/torch-2.4.1%2Bcu124-cp312-cp312-linux_x86_64.whl (797.1 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 797.1/797.1 MB 10.0 MB/s eta 0:00:00
Collecting torchvision==0.19.1
Downloading https://download.pytorch.org/whl/cu124/torchvision-0.19.1%2Bcu124-cp312-cp312-linux_x86_64.whl (7.1 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 7.1/7.1 MB 10.2 MB/s eta 0:00:00
Collecting torchaudio==2.4.1
Downloading https://download.pytorch.org/whl/cu124/torchaudio-2.4.1%2Bcu124-cp312-cp312-linux_x86_64.whl (3.4 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.4/3.4 MB 6.4 MB/s eta 0:00:00
Requirement already satisfied: filelock in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (3.16.1)
Requirement already satisfied: typing-extensions>=4.8.0 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (4.12.2)
Requirement already satisfied: sympy in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (1.13.1)
Requirement already satisfied: networkx in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (3.4.2)
Requirement already satisfied: jinja2 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (3.1.4)
Requirement already satisfied: fsspec in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (2024.10.0)
Requirement already satisfied: setuptools in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (75.2.0)
Collecting nvidia-cuda-nvrtc-cu12==12.4.99 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cuda_nvrtc_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl (24.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 24.7/24.7 MB 8.2 MB/s eta 0:00:00
Collecting nvidia-cuda-runtime-cu12==12.4.99 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cuda_runtime_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl (883 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 883.4/883.4 kB 9.9 MB/s eta 0:00:00
Collecting nvidia-cuda-cupti-cu12==12.4.99 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cuda_cupti_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl (13.8 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 13.8/13.8 MB 10.5 MB/s eta 0:00:00
Requirement already satisfied: nvidia-cudnn-cu12==9.1.0.70 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.1) (9.1.0.70)
Collecting nvidia-cublas-cu12==12.4.2.65 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cublas_cu12-12.4.2.65-py3-none-manylinux2014_x86_64.whl (363.0 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 363.0/363.0 MB 10.1 MB/s eta 0:00:00
Collecting nvidia-cufft-cu12==11.2.0.44 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cufft_cu12-11.2.0.44-py3-none-manylinux2014_x86_64.whl (211.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 211.5/211.5 MB 10.4 MB/s eta 0:00:00
Collecting nvidia-curand-cu12==10.3.5.119 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_curand_cu12-10.3.5.119-py3-none-manylinux2014_x86_64.whl (56.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 56.3/56.3 MB 10.1 MB/s eta 0:00:00
Collecting nvidia-cusolver-cu12==11.6.0.99 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cusolver_cu12-11.6.0.99-py3-none-manylinux2014_x86_64.whl (128.4 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 128.4/128.4 MB 10.4 MB/s eta 0:00:00
Collecting nvidia-cusparse-cu12==12.3.0.142 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_cusparse_cu12-12.3.0.142-py3-none-manylinux2014_x86_64.whl (207.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 207.5/207.5 MB 10.3 MB/s eta 0:00:00
Collecting nvidia-nccl-cu12==2.20.5 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_nccl_cu12-2.20.5-py3-none-manylinux2014_x86_64.whl (176.2 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 176.2/176.2 MB 10.3 MB/s eta 0:00:00
Collecting nvidia-nvtx-cu12==12.4.99 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_nvtx_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl (99 kB)
Collecting nvidia-nvjitlink-cu12==12.4.99 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/cu124/nvidia_nvjitlink_cu12-12.4.99-py3-none-manylinux2014_x86_64.whl (21.1 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 21.1/21.1 MB 10.5 MB/s eta 0:00:00
Collecting triton==3.0.0 (from torch==2.4.1)
Downloading https://download.pytorch.org/whl/triton-3.0.0-1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (209.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 209.5/209.5 MB 10.3 MB/s eta 0:00:00
Requirement already satisfied: numpy in /home/ks/venv/lib/python3.12/site-packages (from torchvision==0.19.1) (1.26.4)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /home/ks/venv/lib/python3.12/site-packages (from torchvision==0.19.1) (11.0.0)
Requirement already satisfied: MarkupSafe>=2.0 in /home/ks/venv/lib/python3.12/site-packages (from jinja2->torch==2.4.1) (3.0.2)
Requirement already satisfied: mpmath<1.4,>=1.1.0 in /home/ks/venv/lib/python3.12/site-packages (from sympy->torch==2.4.1) (1.3.0)
Installing collected packages: triton, nvidia-nvtx-cu12, nvidia-nvjitlink-cu12, nvidia-nccl-cu12, nvidia-curand-cu12, nvidia-cufft-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, nvidia-cublas-cu12, nvidia-cusparse-cu12, nvidia-cusolver-cu12, torch, torchvision, torchaudio
Attempting uninstall: triton
Found existing installation: triton 3.1.0
Uninstalling triton-3.1.0:
Successfully uninstalled triton-3.1.0
Attempting uninstall: nvidia-nvtx-cu12
Found existing installation: nvidia-nvtx-cu12 12.4.127
Uninstalling nvidia-nvtx-cu12-12.4.127:
Successfully uninstalled nvidia-nvtx-cu12-12.4.127
Attempting uninstall: nvidia-nvjitlink-cu12
Found existing installation: nvidia-nvjitlink-cu12 12.4.127
Uninstalling nvidia-nvjitlink-cu12-12.4.127:
Successfully uninstalled nvidia-nvjitlink-cu12-12.4.127
Attempting uninstall: nvidia-nccl-cu12
Found existing installation: nvidia-nccl-cu12 2.21.5
Uninstalling nvidia-nccl-cu12-2.21.5:
Successfully uninstalled nvidia-nccl-cu12-2.21.5
Attempting uninstall: nvidia-curand-cu12
Found existing installation: nvidia-curand-cu12 10.3.5.147
Uninstalling nvidia-curand-cu12-10.3.5.147:
Successfully uninstalled nvidia-curand-cu12-10.3.5.147
Attempting uninstall: nvidia-cufft-cu12
Found existing installation: nvidia-cufft-cu12 11.2.1.3
Uninstalling nvidia-cufft-cu12-11.2.1.3:
Successfully uninstalled nvidia-cufft-cu12-11.2.1.3
Attempting uninstall: nvidia-cuda-runtime-cu12
Found existing installation: nvidia-cuda-runtime-cu12 12.4.127
Uninstalling nvidia-cuda-runtime-cu12-12.4.127:
Successfully uninstalled nvidia-cuda-runtime-cu12-12.4.127
Attempting uninstall: nvidia-cuda-nvrtc-cu12
Found existing installation: nvidia-cuda-nvrtc-cu12 12.4.127
Uninstalling nvidia-cuda-nvrtc-cu12-12.4.127:
Successfully uninstalled nvidia-cuda-nvrtc-cu12-12.4.127
Attempting uninstall: nvidia-cuda-cupti-cu12
Found existing installation: nvidia-cuda-cupti-cu12 12.4.127
Uninstalling nvidia-cuda-cupti-cu12-12.4.127:
Successfully uninstalled nvidia-cuda-cupti-cu12-12.4.127
Attempting uninstall: nvidia-cublas-cu12
Found existing installation: nvidia-cublas-cu12 12.4.5.8
Uninstalling nvidia-cublas-cu12-12.4.5.8:
Successfully uninstalled nvidia-cublas-cu12-12.4.5.8
Attempting uninstall: nvidia-cusparse-cu12
Found existing installation: nvidia-cusparse-cu12 12.3.1.170
Uninstalling nvidia-cusparse-cu12-12.3.1.170:
Successfully uninstalled nvidia-cusparse-cu12-12.3.1.170
Attempting uninstall: nvidia-cusolver-cu12
Found existing installation: nvidia-cusolver-cu12 11.6.1.9
Uninstalling nvidia-cusolver-cu12-11.6.1.9:
Successfully uninstalled nvidia-cusolver-cu12-11.6.1.9
Attempting uninstall: torch
Found existing installation: torch 2.5.0+cu124
Uninstalling torch-2.5.0+cu124:
Successfully uninstalled torch-2.5.0+cu124
Attempting uninstall: torchvision
Found existing installation: torchvision 0.20.0
Uninstalling torchvision-0.20.0:
Successfully uninstalled torchvision-0.20.0
Attempting uninstall: torchaudio
Found existing installation: torchaudio 2.5.0
Uninstalling torchaudio-2.5.0:
Successfully uninstalled torchaudio-2.5.0
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
xformers 0.0.28.post2 requires torch==2.5.0, but you have torch 2.4.1+cu124 which is incompatible.
Successfully installed nvidia-cublas-cu12-12.4.2.65 nvidia-cuda-cupti-cu12-12.4.99 nvidia-cuda-nvrtc-cu12-12.4.99 nvidia-cuda-runtime-cu12-12.4.99 nvidia-cufft-cu12-11.2.0.44 nvidia-curand-cu12-10.3.5.119 nvidia-cusolver-cu12-11.6.0.99 nvidia-cusparse-cu12-12.3.0.142 nvidia-nccl-cu12-2.20.5 nvidia-nvjitlink-cu12-12.4.99 nvidia-nvtx-cu12-12.4.99 torch-2.4.1+cu124 torchaudio-2.4.1+cu124 torchvision-0.19.1+cu124 triton-3.0.0
Buuuuutttttt.... it causes a different error instead...
Restarting... [Legacy Mode]
[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-25 23:04:34.520653
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/venv/bin/python
** ComfyUI Path: /home/ks/venv/comf
** Log path: /home/ks/venv/comf/comfyui.log
Prestartup times for custom nodes:
0.0 seconds: /home/ks/venv/comf/custom_nodes/rgthree-comfy
0.6 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Manager
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.4.1+cu124
WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:
PyTorch 2.5.0+cu124 with CUDA 1204 (you have 2.4.1+cu124)
Python 3.12.7 (you have 3.12.3)
Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers)
Memory-efficient attention, SwiGLU, sparse and more won't be available.
Set XFORMERS_MORE_DETAILS=1 for more details
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using pytorch cross attention
[Prompt Server] web root: /home/ks/venv/comf/web
/home/ks/venv/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
@torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
### Loading: ComfyUI-Manager (V2.51.8)
### ComfyUI Revision: 2785 [52810907] | Released on '2024-10-24'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
/home/ks/venv/lib/python3.12/site-packages/albumentations/__init__.py:24: UserWarning: A new version of Albumentations is available: 1.4.20 (you have 0.0.19). Upgrade using: pip install -U albumentations. To disable automatic update checks, set the environment variable NO_ALBUMENTATIONS_UPDATE to 1.
check_for_updates()
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.4.1+cu124
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.4.1+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
[rgthree-comfy] Loaded 42 fantastic nodes. 🎉
WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/venv/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.
"Art enables us to find ourselves and lose ourselves at the same time." - Thomas Merton
Import times for custom nodes:
0.0 seconds: /home/ks/venv/comf/custom_nodes/websocket_image_save.py
0.0 seconds: /home/ks/venv/comf/custom_nodes/Skimmed_CFG
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI_JPS-Nodes
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-AutomaticCFG
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Image-Saver
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Custom-Scripts
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-KJNodes
0.0 seconds: /home/ks/venv/comf/custom_nodes/rgthree-comfy
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Manager
0.1 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI_ExtraModels
0.3 seconds: /home/ks/venv/comf/custom_nodes/comfyui_face_parsing
0.8 seconds: /home/ks/venv/comf/custom_nodes/comfyui-reactor-node
0.8 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR
1.1 seconds: /home/ks/venv/comf/custom_nodes/was-node-suite-comfyui
Starting server
To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/venv/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
got prompt
model weight dtype torch.float8_e4m3fn, manual cast: torch.float16
model_type EPS
Using pytorch attention in VAE
Using pytorch attention in VAE
Model maximum sigma: 14.614640235900879 / Model minimum sigma: 0.029167160391807556
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 2448.5241737365723 True
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████| 100/100 [00:38<00:00, 2.59it/s]
Requested to load AutoencoderKL
Loading 1 new model
loaded completely 0.0 159.55708122253418 True
model weight dtype torch.float16, manual cast: None
model_type EPS
Using pytorch attention in VAE
Using pytorch attention in VAE
Model maximum sigma: 14.614640235900879 / Model minimum sigma: 0.029167160391807556
Loading weights to: cpu
Diffusion using fp16
making attention of type 'vanilla' with 512 in_channels
Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
making attention of type 'vanilla' with 512 in_channels
Attempting to load SDXL model from node inputs
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
Loading first clip model from SDXL checkpoint
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Loading second clip model from SDXL checkpoint
Attempting to load SUPIR model: [/home/ks/venv/comf/models/checkpoints/SUPIR_pruned/SUPIR-v0Q_fp16.safetensors]
Loaded state_dict from [/home/ks/venv/comf/models/checkpoints/SUPIR_pruned/SUPIR-v0Q_fp16.safetensors]
Encoder using bf16
[Tiled VAE]: input_size: torch.Size([1, 3, 2688, 1536]), tile_size: 512, padding: 32
[Tiled VAE]: split to 6x3 = 18 tiles. Optimal tile size 512x448, original tile size 512x512
[Tiled VAE]: Executing Encoder Task Queue: 100%|████████████████████████████████████████████████████████| 1638/1638 [00:06<00:00, 235.35it/s]
[Tiled VAE]: Done in 7.515s, max VRAM alloc 664.499 MB
[Tiled VAE]: input_size: torch.Size([1, 4, 336, 192]), tile_size: 64, padding: 11
[Tiled VAE]: split to 5x3 = 15 tiles. Optimal tile size 64x64, original tile size 64x64
[Tiled VAE]: Executing Decoder Task Queue: 100%|████████████████████████████████████████████████████████| 1845/1845 [00:12<00:00, 142.10it/s]
[Tiled VAE]: Done in 13.520s, max VRAM alloc 1760.142 MB
captions: [['']]
Batch captioning
Encoder using bf16
[Tiled VAE]: input_size: torch.Size([1, 3, 2688, 1536]), tile_size: 512, padding: 32
[Tiled VAE]: split to 6x3 = 18 tiles. Optimal tile size 512x448, original tile size 512x512
[Tiled VAE]: Executing Encoder Task Queue: 100%|████████████████████████████████████████████████████████| 1638/1638 [00:04<00:00, 352.28it/s]
[Tiled VAE]: Done in 5.209s, max VRAM alloc 667.091 MB
sampler_config: {'target': '.sgm.modules.diffusionmodules.sampling.RestoreDPMPP2MSampler', 'params': {'num_steps': 16, 'restore_cfg': -1, 's_churn': 5, 's_noise': 1.0030000000000001, 'discretization_config': {'target': '.sgm.modules.diffusionmodules.discretizer.LegacyDDPMDiscretization'}, 'guider_config': {'target': '.sgm.modules.diffusionmodules.guiders.LinearCFG', 'params': {'scale': 4.0, 'scale_min': 4.0}}, 'eta': 1.0}}
Using random noise
!!! Exception during processing !!! No operator found for `memory_efficient_attention_forward` with inputs:
query : shape=(40, 4032, 1, 64) (torch.float16)
key : shape=(40, 4032, 1, 64) (torch.float16)
value : shape=(40, 4032, 1, 64) (torch.float16)
attn_bias : <class 'NoneType'>
p : 0.0
`fa2F@v2.5.6-pt` is not supported because:
xFormers wasn't build with CUDA support
`cutlassF` is not supported because:
xFormers wasn't build with CUDA support
operator wasn't built - see `python -m xformers.info` for more info
Traceback (most recent call last):
File "/home/ks/venv/comf/execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "/home/ks/venv/comf/execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/nodes_v2.py", line 504, in sample
_samples = self.sampler(denoiser, noised_z, cond=positive[i], uc=negative[i], x_center=sample.unsqueeze(0), control_scale=control_scale_end,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/sgm/modules/diffusionmodules/sampling.py", line 657, in __call__
x, old_denoised = self.sampler_step(
^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/sgm/modules/diffusionmodules/sampling.py", line 609, in sampler_step
denoised = self.denoise(x, denoiser, sigma, cond, uc, control_scale=control_scale)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/sgm/modules/diffusionmodules/sampling.py", line 573, in denoise
denoised = denoiser(*self.guider.prepare_inputs(x, sigma, cond, uc), control_scale)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/nodes_v2.py", line 477, in <lambda>
denoiser = lambda input, sigma, c, control_scale: SUPIR_model.denoiser(SUPIR_model.model, input, sigma, c, control_scale)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/sgm/modules/diffusionmodules/denoiser.py", line 73, in __call__
return network(input * c_in, c_noise, cond, control_scale) * c_out + input * c_skip
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/sgm/modules/diffusionmodules/wrappers.py", line 96, in forward
out = self.diffusion_model(
^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/SUPIR/modules/SUPIR_v0.py", line 657, in forward
h = self.project_modules[adapter_idx](control[control_idx], h, control_scale=control_scale)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/SUPIR/modules/SUPIR_v0.py", line 150, in forward
x = self.attn(x, context)
^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR/sgm/modules/attention.py", line 360, in forward
out = xformers.ops.memory_efficient_attention(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/__init__.py", line 301, in memory_efficient_attention
return _memory_efficient_attention(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/__init__.py", line 462, in _memory_efficient_attention
return _memory_efficient_attention_forward(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/__init__.py", line 481, in _memory_efficient_attention_forward
op = _dispatch_fw(inp, False)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/dispatch.py", line 135, in _dispatch_fw
return _run_priority_list(
^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/dispatch.py", line 76, in _run_priority_list
raise NotImplementedError(msg)
NotImplementedError: No operator found for `memory_efficient_attention_forward` with inputs:
query : shape=(40, 4032, 1, 64) (torch.float16)
key : shape=(40, 4032, 1, 64) (torch.float16)
value : shape=(40, 4032, 1, 64) (torch.float16)
attn_bias : <class 'NoneType'>
p : 0.0
`fa2F@v2.5.6-pt` is not supported because:
xFormers wasn't build with CUDA support
`cutlassF` is not supported because:
xFormers wasn't build with CUDA support
operator wasn't built - see `python -m xformers.info` for more info
Prompt executed in 142.48 seconds
so i thought... okay... so something is wrong with the xformers... maybe i should go back an older version...
so i did...
on bash ks 3.12.3 master ≡ 3m 56s 161ms
╭─ 23:02:43 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers <=0.028
-bash: =0.028: No such file or directory
on bash ks 3.12.3 master ≡ 5ms
╭─ 23:09:50 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers =<0.028
-bash: 0.028: No such file or directory
on bash ks 3.12.3 master ≡ 4ms
╭─ 23:09:58 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers =<0.0.28
-bash: 0.0.28: No such file or directory
on bash ks 3.12.3 master ≡ 5ms
╭─ 23:10:08 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers 0.0.27
Requirement already satisfied: xformers in /home/ks/venv/lib/python3.12/site-packages (0.0.28.post2)
ERROR: Could not find a version that satisfies the requirement 0.0.27 (from versions: none)
ERROR: No matching distribution found for 0.0.27
on bash ks 3.12.3 master ≡ 8s 536ms
╭─ 23:10:26 | 25 Oct, Friday | in venv comf
╰─❯ pip --help
Usage:
pip <command> [options]
Commands:
install Install packages.
download Download packages.
uninstall Uninstall packages.
freeze Output installed packages in requirements format.
inspect Inspect the python environment.
list List installed packages.
show Show information about installed packages.
check Verify installed packages have compatible dependencies.
config Manage local and global configuration.
search Search PyPI for packages.
cache Inspect and manage pip's wheel cache.
index Inspect information available from package indexes.
wheel Build wheels from your requirements.
hash Compute hashes of package archives.
completion A helper command used for command completion.
debug Show information useful for debugging.
help Show help for commands.
General Options:
-h, --help Show help.
--debug Let unhandled exceptions propagate outside the main subroutine, instead of logging them to stderr.
--isolated Run pip in an isolated mode, ignoring environment variables and user configuration.
--require-virtualenv Allow pip to only run in a virtual environment; exit with an error otherwise.
--python <python> Run pip with the specified Python interpreter.
-v, --verbose Give more output. Option is additive, and can be used up to 3 times.
-V, --version Show version and exit.
-q, --quiet Give less output. Option is additive, and can be used up to 3 times (corresponding to WARNING, ERROR, and
CRITICAL logging levels).
--log <path> Path to a verbose appending log.
--no-input Disable prompting for input.
--keyring-provider <keyring_provider>
Enable the credential lookup via the keyring library if user input is allowed. Specify which mechanism to use
[disabled, import, subprocess]. (default: disabled)
--proxy <proxy> Specify a proxy in the form scheme://[user:passwd@]proxy.server:port.
--retries <retries> Maximum number of retries each connection should attempt (default 5 times).
--timeout <sec> Set the socket timeout (default 15 seconds).
--exists-action <action> Default action when a path already exists: (s)witch, (i)gnore, (w)ipe, (b)ackup, (a)bort.
--trusted-host <hostname> Mark this host or host:port pair as trusted, even though it does not have valid or any HTTPS.
--cert <path> Path to PEM-encoded CA certificate bundle. If provided, overrides the default. See 'SSL Certificate
Verification' in pip documentation for more information.
--client-cert <path> Path to SSL client certificate, a single file containing the private key and the certificate in PEM format.
--cache-dir <dir> Store the cache data in <dir>.
--no-cache-dir Disable the cache.
--disable-pip-version-check
Don't periodically check PyPI to determine whether a new version of pip is available for download. Implied
with --no-index.
--no-color Suppress colored output.
--no-python-version-warning
Silence deprecation warnings for upcoming unsupported Pythons.
--use-feature <feature> Enable new functionality, that may be backward incompatible.
--use-deprecated <feature> Enable deprecated functionality, that will be removed in the future.
on bash ks 3.12.3 master ≡ 7s 949ms
╭─ 23:10:41 | 25 Oct, Friday | in venv comf
╰─❯ pip uninstall xformers
Found existing installation: xformers 0.0.28.post2
Uninstalling xformers-0.0.28.post2:
Would remove:
/home/ks/venv/lib/python3.12/site-packages/xformers-0.0.28.post2.dist-info/*
/home/ks/venv/lib/python3.12/site-packages/xformers/*
Proceed (Y/n)? y
Successfully uninstalled xformers-0.0.28.post2
on bash ks 3.12.3 master ≡ 4s 41ms
╭─ 23:11:01 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers 0.0.27
Collecting xformers
Downloading xformers-0.0.28.post2-cp312-cp312-manylinux_2_28_x86_64.whl.metadata (1.0 kB)
ERROR: Could not find a version that satisfies the requirement 0.0.27 (from versions: none)
ERROR: No matching distribution found for 0.0.27
on bash ks 3.12.3 master ≡ 477ms
╭─ 23:11:13 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers <0.0.28
-bash: 0.0.28: No such file or directory
on bash ks 3.12.3 master ≡ 4ms
╭─ 23:11:38 | 25 Oct, Friday | in venv comf
╰─❯ pip install xformers==0.0.27.post2
Collecting xformers==0.0.27.post2
Downloading xformers-0.0.27.post2-cp312-cp312-manylinux2014_x86_64.whl.metadata (1.0 kB)
Requirement already satisfied: numpy in /home/ks/venv/lib/python3.12/site-packages (from xformers==0.0.27.post2) (1.26.4)
Collecting torch==2.4.0 (from xformers==0.0.27.post2)
Downloading torch-2.4.0-cp312-cp312-manylinux1_x86_64.whl.metadata (26 kB)
Requirement already satisfied: filelock in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (3.16.1)
Requirement already satisfied: typing-extensions>=4.8.0 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (4.12.2)
Requirement already satisfied: sympy in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (1.13.1)
Requirement already satisfied: networkx in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (3.4.2)
Requirement already satisfied: jinja2 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (3.1.4)
Requirement already satisfied: fsspec in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (2024.10.0)
Requirement already satisfied: setuptools in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (75.2.0)
Collecting nvidia-cuda-nvrtc-cu12==12.1.105 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cuda_nvrtc_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cuda-runtime-cu12==12.1.105 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cuda_runtime_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cuda-cupti-cu12==12.1.105 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cuda_cupti_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Requirement already satisfied: nvidia-cudnn-cu12==9.1.0.70 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (9.1.0.70)
Collecting nvidia-cublas-cu12==12.1.3.1 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cublas_cu12-12.1.3.1-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cufft-cu12==11.0.2.54 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cufft_cu12-11.0.2.54-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-curand-cu12==10.3.2.106 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_curand_cu12-10.3.2.106-py3-none-manylinux1_x86_64.whl.metadata (1.5 kB)
Collecting nvidia-cusolver-cu12==11.4.5.107 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cusolver_cu12-11.4.5.107-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Collecting nvidia-cusparse-cu12==12.1.0.106 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_cusparse_cu12-12.1.0.106-py3-none-manylinux1_x86_64.whl.metadata (1.6 kB)
Requirement already satisfied: nvidia-nccl-cu12==2.20.5 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (2.20.5)
Collecting nvidia-nvtx-cu12==12.1.105 (from torch==2.4.0->xformers==0.0.27.post2)
Using cached nvidia_nvtx_cu12-12.1.105-py3-none-manylinux1_x86_64.whl.metadata (1.7 kB)
Requirement already satisfied: triton==3.0.0 in /home/ks/venv/lib/python3.12/site-packages (from torch==2.4.0->xformers==0.0.27.post2) (3.0.0)
Requirement already satisfied: nvidia-nvjitlink-cu12 in /home/ks/venv/lib/python3.12/site-packages (from nvidia-cusolver-cu12==11.4.5.107->torch==2.4.0->xformers==0.0.27.post2) (12.4.99)
Requirement already satisfied: MarkupSafe>=2.0 in /home/ks/venv/lib/python3.12/site-packages (from jinja2->torch==2.4.0->xformers==0.0.27.post2) (3.0.2)
Requirement already satisfied: mpmath<1.4,>=1.1.0 in /home/ks/venv/lib/python3.12/site-packages (from sympy->torch==2.4.0->xformers==0.0.27.post2) (1.3.0)
Downloading xformers-0.0.27.post2-cp312-cp312-manylinux2014_x86_64.whl (20.8 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 20.8/20.8 MB 10.5 MB/s eta 0:00:00
Downloading torch-2.4.0-cp312-cp312-manylinux1_x86_64.whl (797.2 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 797.2/797.2 MB 10.1 MB/s eta 0:00:00
Using cached nvidia_cublas_cu12-12.1.3.1-py3-none-manylinux1_x86_64.whl (410.6 MB)
Using cached nvidia_cuda_cupti_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (14.1 MB)
Using cached nvidia_cuda_nvrtc_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (23.7 MB)
Using cached nvidia_cuda_runtime_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (823 kB)
Using cached nvidia_cufft_cu12-11.0.2.54-py3-none-manylinux1_x86_64.whl (121.6 MB)
Using cached nvidia_curand_cu12-10.3.2.106-py3-none-manylinux1_x86_64.whl (56.5 MB)
Using cached nvidia_cusolver_cu12-11.4.5.107-py3-none-manylinux1_x86_64.whl (124.2 MB)
Using cached nvidia_cusparse_cu12-12.1.0.106-py3-none-manylinux1_x86_64.whl (196.0 MB)
Using cached nvidia_nvtx_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (99 kB)
Installing collected packages: nvidia-nvtx-cu12, nvidia-cusparse-cu12, nvidia-curand-cu12, nvidia-cufft-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, nvidia-cublas-cu12, nvidia-cusolver-cu12, torch, xformers
Attempting uninstall: nvidia-nvtx-cu12
Found existing installation: nvidia-nvtx-cu12 12.4.99
Uninstalling nvidia-nvtx-cu12-12.4.99:
Successfully uninstalled nvidia-nvtx-cu12-12.4.99
Attempting uninstall: nvidia-cusparse-cu12
Found existing installation: nvidia-cusparse-cu12 12.3.0.142
Uninstalling nvidia-cusparse-cu12-12.3.0.142:
Successfully uninstalled nvidia-cusparse-cu12-12.3.0.142
Attempting uninstall: nvidia-curand-cu12
Found existing installation: nvidia-curand-cu12 10.3.5.119
Uninstalling nvidia-curand-cu12-10.3.5.119:
Successfully uninstalled nvidia-curand-cu12-10.3.5.119
Attempting uninstall: nvidia-cufft-cu12
Found existing installation: nvidia-cufft-cu12 11.2.0.44
Uninstalling nvidia-cufft-cu12-11.2.0.44:
Successfully uninstalled nvidia-cufft-cu12-11.2.0.44
Attempting uninstall: nvidia-cuda-runtime-cu12
Found existing installation: nvidia-cuda-runtime-cu12 12.4.99
Uninstalling nvidia-cuda-runtime-cu12-12.4.99:
Successfully uninstalled nvidia-cuda-runtime-cu12-12.4.99
Attempting uninstall: nvidia-cuda-nvrtc-cu12
Found existing installation: nvidia-cuda-nvrtc-cu12 12.4.99
Uninstalling nvidia-cuda-nvrtc-cu12-12.4.99:
Successfully uninstalled nvidia-cuda-nvrtc-cu12-12.4.99
Attempting uninstall: nvidia-cuda-cupti-cu12
Found existing installation: nvidia-cuda-cupti-cu12 12.4.99
Uninstalling nvidia-cuda-cupti-cu12-12.4.99:
Successfully uninstalled nvidia-cuda-cupti-cu12-12.4.99
Attempting uninstall: nvidia-cublas-cu12
Found existing installation: nvidia-cublas-cu12 12.4.2.65
Uninstalling nvidia-cublas-cu12-12.4.2.65:
Successfully uninstalled nvidia-cublas-cu12-12.4.2.65
Attempting uninstall: nvidia-cusolver-cu12
Found existing installation: nvidia-cusolver-cu12 11.6.0.99
Uninstalling nvidia-cusolver-cu12-11.6.0.99:
Successfully uninstalled nvidia-cusolver-cu12-11.6.0.99
Attempting uninstall: torch
Found existing installation: torch 2.4.1+cu124
Uninstalling torch-2.4.1+cu124:
Successfully uninstalled torch-2.4.1+cu124
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
torchaudio 2.4.1+cu124 requires torch==2.4.1, but you have torch 2.4.0 which is incompatible.
torchvision 0.19.1+cu124 requires torch==2.4.1, but you have torch 2.4.0 which is incompatible.
Successfully installed nvidia-cublas-cu12-12.1.3.1 nvidia-cuda-cupti-cu12-12.1.105 nvidia-cuda-nvrtc-cu12-12.1.105 nvidia-cuda-runtime-cu12-12.1.105 nvidia-cufft-cu12-11.0.2.54 nvidia-curand-cu12-10.3.2.106 nvidia-cusolver-cu12-11.4.5.107 nvidia-cusparse-cu12-12.1.0.106 nvidia-nvtx-cu12-12.1.105 torch-2.4.0 xformers-0.0.27.post2
on bash ks 3.12.3 master ≡ 1m 33s 481ms
╭─ 23:14:12 | 25 Oct, Friday | in venv comf
╰─❯
now i get this annoying error...
but it is related to audio... which means... i can live with it for now.
and my SUPIR workflow completed.
Restarting... [Legacy Mode]
[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-25 23:14:33.237226
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/venv/bin/python
** ComfyUI Path: /home/ks/venv/comf
** Log path: /home/ks/venv/comf/comfyui.log
Prestartup times for custom nodes:
0.0 seconds: /home/ks/venv/comf/custom_nodes/rgthree-comfy
0.5 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Manager
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.4.0+cu121
/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/flash.py:211: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch.
@torch.library.impl_abstract("xformers_flash::flash_fwd")
/home/ks/venv/lib/python3.12/site-packages/xformers/ops/fmha/flash.py:344: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch.
@torch.library.impl_abstract("xformers_flash::flash_bwd")
xformers version: 0.0.27.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using xformers cross attention
[Prompt Server] web root: /home/ks/venv/comf/web
/home/ks/venv/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
@torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
Traceback (most recent call last):
File "/home/ks/venv/comf/nodes.py", line 2001, in load_custom_node
module_spec.loader.exec_module(module)
File "<frozen importlib._bootstrap_external>", line 995, in exec_module
File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
File "/home/ks/venv/comf/comfy_extras/nodes_audio.py", line 1, in <module>
import torchaudio
File "/home/ks/venv/lib/python3.12/site-packages/torchaudio/__init__.py", line 2, in <module>
from . import _extension # noqa # usort: skip
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/ks/venv/lib/python3.12/site-packages/torchaudio/_extension/__init__.py", line 42, in <module>
_check_cuda_version()
File "/home/ks/venv/lib/python3.12/site-packages/torchaudio/_extension/utils.py", line 175, in _check_cuda_version
raise RuntimeError(
RuntimeError: Detected that PyTorch and TorchAudio were compiled with different CUDA versions. PyTorch has CUDA version 12.1 whereas TorchAudio has CUDA version 12.4. Please install the TorchAudio version that matches your PyTorch version.
Cannot import /home/ks/venv/comf/comfy_extras/nodes_audio.py module for custom nodes: Detected that PyTorch and TorchAudio were compiled with different CUDA versions. PyTorch has CUDA version 12.1 whereas TorchAudio has CUDA version 12.4. Please install the TorchAudio version that matches your PyTorch version.
### Loading: ComfyUI-Manager (V2.51.8)
### ComfyUI Revision: 2785 [52810907] | Released on '2024-10-24'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
/home/ks/venv/lib/python3.12/site-packages/albumentations/__init__.py:24: UserWarning: A new version of Albumentations is available: 1.4.20 (you have 0.0.19). Upgrade using: pip install -U albumentations. To disable automatic update checks, set the environment variable NO_ALBUMENTATIONS_UPDATE to 1.
check_for_updates()
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.4.0+cu121
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.4.0+cu121
xformers version: 0.0.27.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
[rgthree-comfy] Loaded 42 epic nodes. 🎉
WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/venv/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.
"Art is the journey of a free soul." - Alev Oguz
Import times for custom nodes:
0.0 seconds: /home/ks/venv/comf/custom_nodes/websocket_image_save.py
0.0 seconds: /home/ks/venv/comf/custom_nodes/Skimmed_CFG
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI_JPS-Nodes
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-AutomaticCFG
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Image-Saver
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Custom-Scripts
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-KJNodes
0.0 seconds: /home/ks/venv/comf/custom_nodes/rgthree-comfy
0.0 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-Manager
0.1 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI_ExtraModels
0.2 seconds: /home/ks/venv/comf/custom_nodes/comfyui_face_parsing
0.8 seconds: /home/ks/venv/comf/custom_nodes/ComfyUI-SUPIR
0.9 seconds: /home/ks/venv/comf/custom_nodes/comfyui-reactor-node
1.0 seconds: /home/ks/venv/comf/custom_nodes/was-node-suite-comfyui
WARNING: some comfy_extras/ nodes did not import correctly. This may be because they are missing some dependencies.
IMPORT FAILED: nodes_audio.py
This issue might be caused by new missing dependencies added the last time you updated ComfyUI.
Please do a: pip install -r requirements.txt
Starting server
To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/venv/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
got prompt
model weight dtype torch.float8_e4m3fn, manual cast: torch.float16
model_type EPS
Using xformers attention in VAE
Using xformers attention in VAE
Model maximum sigma: 14.614640235900879 / Model minimum sigma: 0.029167160391807556
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 2448.5241737365723 True
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████| 100/100 [00:40<00:00, 2.46it/s]
Requested to load AutoencoderKL
Loading 1 new model
loaded completely 0.0 159.55708122253418 True
model weight dtype torch.float16, manual cast: None
model_type EPS
Using xformers attention in VAE
Using xformers attention in VAE
Model maximum sigma: 14.614640235900879 / Model minimum sigma: 0.029167160391807556
Loading weights to: cpu
Diffusion using fp16
Using XFORMERS
making attention of type 'vanilla-xformers' with 512 in_channels
building MemoryEfficientAttnBlock with 512 in_channels...
Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
making attention of type 'vanilla-xformers' with 512 in_channels
building MemoryEfficientAttnBlock with 512 in_channels...
Attempting to load SDXL model from node inputs
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
Loading first clip model from SDXL checkpoint
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Loading second clip model from SDXL checkpoint
Attempting to load SUPIR model: [/home/ks/venv/comf/models/checkpoints/SUPIR_pruned/SUPIR-v0Q_fp16.safetensors]
Loaded state_dict from [/home/ks/venv/comf/models/checkpoints/SUPIR_pruned/SUPIR-v0Q_fp16.safetensors]
Encoder using bf16
[Tiled VAE]: input_size: torch.Size([1, 3, 2688, 1536]), tile_size: 512, padding: 32
[Tiled VAE]: split to 6x3 = 18 tiles. Optimal tile size 512x448, original tile size 512x512
[Tiled VAE]: Executing Encoder Task Queue: 100%|████████████████████████████████████████████████████████| 1638/1638 [00:06<00:00, 245.81it/s]
[Tiled VAE]: Done in 7.213s, max VRAM alloc 664.499 MB
[Tiled VAE]: input_size: torch.Size([1, 4, 336, 192]), tile_size: 64, padding: 11
[Tiled VAE]: split to 5x3 = 15 tiles. Optimal tile size 64x64, original tile size 64x64
[Tiled VAE]: Executing Decoder Task Queue: 100%|████████████████████████████████████████████████████████| 1845/1845 [00:12<00:00, 146.60it/s]
[Tiled VAE]: Done in 13.116s, max VRAM alloc 1760.142 MB
captions: [['']]
Batch captioning
Encoder using bf16
[Tiled VAE]: input_size: torch.Size([1, 3, 2688, 1536]), tile_size: 512, padding: 32
[Tiled VAE]: split to 6x3 = 18 tiles. Optimal tile size 512x448, original tile size 512x512
[Tiled VAE]: Executing Encoder Task Queue: 100%|████████████████████████████████████████████████████████| 1638/1638 [00:04<00:00, 362.45it/s]
[Tiled VAE]: Done in 5.071s, max VRAM alloc 667.091 MB
sampler_config: {'target': '.sgm.modules.diffusionmodules.sampling.RestoreDPMPP2MSampler', 'params': {'num_steps': 16, 'restore_cfg': -1, 's_churn': 5, 's_noise': 1.0030000000000001, 'discretization_config': {'target': '.sgm.modules.diffusionmodules.discretizer.LegacyDDPMDiscretization'}, 'guider_config': {'target': '.sgm.modules.diffusionmodules.guiders.LinearCFG', 'params': {'scale': 4.0, 'scale_min': 4.0}}, 'eta': 1.0}}
Using random noise
Sampled 1 of 1
Decoder using bf16
SUPIR decoder using torch.bfloat16
[Tiled VAE]: input_size: torch.Size([1, 4, 336, 192]), tile_size: 64, padding: 11
[Tiled VAE]: split to 5x3 = 15 tiles. Optimal tile size 64x64, original tile size 64x64
[Tiled VAE]: Executing Decoder Task Queue: 100%|████████████████████████████████████████████████████████| 1845/1845 [00:10<00:00, 169.13it/s]
[Tiled VAE]: Done in 11.450s, max VRAM alloc 5398.604 MB
Prompt executed in 178.14 seconds
so on one hand... i get to generate my images with the workflow i planned on using... but on the other hand... i downgraded my torch and xformers to an older version which makes me feel yucky.
does anyone have any recommendation?
just live with it until the next thing that breaks?
lol.
Your question
okay so below is the first error that i encountered when trying to use a SUPIR workflow.
the subsequent steps are what i used to get my comfyui running but with other errors...
Logs
Other
when searching online this is the first item that i found similar to what i was facing:
https://github.com/huggingface/diffusers/issues/9704#issuecomment-2423400790