Open BenDes21 opened 1 month ago
same thing happened to me :(
same
same
same
I'm experiencing something similar. Every time I change the text prompt. Once I've run the prompt I can do it again successfully. But it fails every first try. Here's the log
Exception Message: Allocation on device
File "/workspace/ComfyUI/execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "/workspace/ComfyUI/execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "/workspace/ComfyUI/execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
File "/workspace/ComfyUI/nodes.py", line 65, in encode
output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True)
File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens
o = self.cond_stage_model.encode_token_weights(tokens)
File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights
t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5)
File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights
o = self.encode(to_encode)
File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode
return self(tokens)
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward
outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32)
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward
x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32))
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
return forward_call(*args, **kwargs)
File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward
return self.forward_comfy_cast_weights(*args, **kwargs)
File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights
return self.forward_ggml_cast_weights(input, *args, **kwargs)
File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights
weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype)
File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight
weight = s.get_weight(s.weight.to(device), dtype)
File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight
weight = dequantize_tensor(tensor, dtype, self.dequant_dtype)
File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor
return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype)
File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to
new = super().to(*args, **kwargs)
File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__
ret = func(*args, **kwargs)
## System Information
- **ComfyUI Version:** v0.2.3-8-gf584758
- **Arguments:** main.py --disable-auto-launch --port 18188
- **OS:** posix
- **Python Version:** 3.10.12 (main, Sep 11 2024, 15:47:36) [GCC 11.4.0]
- **Embedded Python:** false
- **PyTorch Version:** 2.4.1+cu121
## Devices
- **Name:** cuda:0 NVIDIA RTX 6000 Ada Generation : cudaMallocAsync
- **Type:** cuda
- **VRAM Total:** 51010207744
- **VRAM Free:** 18704695296
- **Torch VRAM Total:** 67108864
- **Torch VRAM Free:** 58589184
## Logs
2024-10-15 19:01:34,134 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 19:01:34,137 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 19:01:36,154 - root - INFO - Prompt executed in 2.54 seconds 2024-10-15 19:01:40,556 - root - INFO - got prompt 2024-10-15 19:01:40,573 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:01:40,573 - root - INFO - Loading 1 new model 2024-10-15 19:01:41,145 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:01:41,310 - root - INFO - Requested to load Flux 2024-10-15 19:01:41,310 - root - INFO - Loading 1 new model 2024-10-15 19:01:43,011 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 19:01:44,551 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 19:01:44,551 - root - INFO - Loading 1 new model 2024-10-15 19:01:44,579 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 19:01:44,781 - root - INFO - Prompt executed in 4.21 seconds 2024-10-15 19:01:53,811 - root - INFO - got prompt 2024-10-15 19:01:55,554 - root - INFO - Prompt executed in 1.73 seconds 2024-10-15 19:02:02,650 - root - INFO - got prompt 2024-10-15 19:02:04,400 - root - INFO - Prompt executed in 1.74 seconds 2024-10-15 19:02:15,253 - root - INFO - got prompt 2024-10-15 19:02:17,017 - root - INFO - Prompt executed in 1.75 seconds 2024-10-15 19:02:24,659 - root - INFO - got prompt 2024-10-15 19:02:26,440 - root - INFO - Prompt executed in 1.77 seconds 2024-10-15 19:02:37,626 - root - INFO - got prompt 2024-10-15 19:02:39,396 - root - INFO - Prompt executed in 1.76 seconds 2024-10-15 19:02:54,700 - root - INFO - got prompt 2024-10-15 19:02:54,717 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:02:54,717 - root - INFO - Loading 1 new model 2024-10-15 19:02:55,235 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:02:55,250 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 19:02:55,256 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 19:02:55,259 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 19:02:57,615 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 19:03:01,228 - root - INFO - got prompt 2024-10-15 19:03:01,265 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:03:01,266 - root - INFO - Loading 1 new model 2024-10-15 19:03:01,687 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:03:01,854 - root - INFO - Requested to load Flux 2024-10-15 19:03:01,855 - root - INFO - Loading 1 new model 2024-10-15 19:03:03,512 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 19:03:05,067 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 19:03:05,067 - root - INFO - Loading 1 new model 2024-10-15 19:03:05,094 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 19:03:05,302 - root - INFO - Prompt executed in 4.04 seconds 2024-10-15 19:03:14,582 - root - INFO - got prompt 2024-10-15 19:03:16,364 - root - INFO - Prompt executed in 1.77 seconds 2024-10-15 19:03:22,869 - root - INFO - got prompt 2024-10-15 19:03:24,627 - root - INFO - Prompt executed in 1.75 seconds 2024-10-15 19:03:34,347 - root - INFO - got prompt 2024-10-15 19:03:34,365 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:03:34,365 - root - INFO - Loading 1 new model 2024-10-15 19:03:34,784 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:03:34,799 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 19:03:34,804 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 19:03:34,806 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 19:03:36,957 - root - INFO - Prompt executed in 2.60 seconds 2024-10-15 19:03:40,405 - root - INFO - got prompt 2024-10-15 19:03:40,428 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:03:40,429 - root - INFO - Loading 1 new model 2024-10-15 19:03:40,859 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:03:41,029 - root - INFO - Requested to load Flux 2024-10-15 19:03:41,030 - root - INFO - Loading 1 new model 2024-10-15 19:03:42,636 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 19:03:44,181 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 19:03:44,181 - root - INFO - Loading 1 new model 2024-10-15 19:03:44,238 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 19:03:44,431 - root - INFO - Prompt executed in 4.01 seconds 2024-10-15 19:03:49,936 - root - INFO - got prompt 2024-10-15 19:03:51,700 - root - INFO - Prompt executed in 1.75 seconds 2024-10-15 19:04:03,688 - root - INFO - got prompt 2024-10-15 19:04:03,705 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:04:03,705 - root - INFO - Loading 1 new model 2024-10-15 19:04:04,410 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:04:04,425 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 19:04:04,430 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 19:04:04,433 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 19:04:06,336 - root - INFO - Prompt executed in 2.63 seconds 2024-10-15 19:04:09,723 - root - INFO - got prompt 2024-10-15 19:04:09,737 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:04:09,737 - root - INFO - Loading 1 new model 2024-10-15 19:04:10,147 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:04:10,311 - root - INFO - Requested to load Flux 2024-10-15 19:04:10,312 - root - INFO - Loading 1 new model 2024-10-15 19:04:11,795 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 19:04:13,348 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 19:04:13,349 - root - INFO - Loading 1 new model 2024-10-15 19:04:13,378 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 19:04:13,578 - root - INFO - Prompt executed in 3.84 seconds 2024-10-15 19:04:17,539 - root - INFO - got prompt 2024-10-15 19:04:19,335 - root - INFO - Prompt executed in 1.78 seconds 2024-10-15 19:04:23,699 - root - INFO - got prompt 2024-10-15 19:04:25,500 - root - INFO - Prompt executed in 1.79 seconds 2024-10-15 19:04:32,345 - root - INFO - got prompt 2024-10-15 19:04:34,135 - root - INFO - Prompt executed in 1.78 seconds 2024-10-15 19:04:38,334 - root - INFO - got prompt 2024-10-15 19:04:40,166 - root - INFO - Prompt executed in 1.82 seconds 2024-10-15 19:04:45,344 - root - INFO - got prompt 2024-10-15 19:04:47,113 - root - INFO - Prompt executed in 1.76 seconds 2024-10-15 19:04:51,646 - root - INFO - got prompt 2024-10-15 19:04:53,418 - root - INFO - Prompt executed in 1.76 seconds 2024-10-15 19:04:55,717 - root - INFO - got prompt 2024-10-15 19:04:57,538 - root - INFO - Prompt executed in 1.81 seconds 2024-10-15 19:05:07,540 - root - INFO - got prompt 2024-10-15 19:05:09,328 - root - INFO - Prompt executed in 1.78 seconds 2024-10-15 19:05:15,892 - root - INFO - got prompt 2024-10-15 19:05:17,668 - root - INFO - Prompt executed in 1.76 seconds 2024-10-15 19:05:38,901 - root - INFO - got prompt 2024-10-15 19:05:38,918 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:05:38,918 - root - INFO - Loading 1 new model 2024-10-15 19:05:40,110 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:05:40,124 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 19:05:40,128 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 19:05:40,130 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 19:05:42,529 - root - INFO - Prompt executed in 3.62 seconds 2024-10-15 19:05:46,675 - root - INFO - got prompt 2024-10-15 19:05:46,689 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:05:46,689 - root - INFO - Loading 1 new model 2024-10-15 19:05:47,114 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:05:47,281 - root - INFO - Requested to load Flux 2024-10-15 19:05:47,282 - root - INFO - Loading 1 new model 2024-10-15 19:05:48,947 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 19:05:50,507 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 19:05:50,507 - root - INFO - Loading 1 new model 2024-10-15 19:05:50,536 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 19:05:50,757 - root - INFO - Prompt executed in 4.07 seconds 2024-10-15 19:05:56,814 - root - INFO - got prompt 2024-10-15 19:05:58,592 - root - INFO - Prompt executed in 1.77 seconds 2024-10-15 19:06:01,360 - root - INFO - got prompt 2024-10-15 19:06:03,196 - root - INFO - Prompt executed in 1.83 seconds 2024-10-15 19:06:37,558 - root - INFO - got prompt 2024-10-15 19:06:37,572 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:06:37,572 - root - INFO - Loading 1 new model 2024-10-15 19:06:37,963 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:06:37,978 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 19:06:37,983 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 19:06:37,985 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 19:06:40,145 - root - INFO - Prompt executed in 2.58 seconds 2024-10-15 19:06:42,671 - root - INFO - got prompt 2024-10-15 19:06:42,686 - root - INFO - Requested to load FluxClipModel 2024-10-15 19:06:42,686 - root - INFO - Loading 1 new model 2024-10-15 19:06:43,065 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 19:06:43,234 - root - INFO - Requested to load Flux 2024-10-15 19:06:43,235 - root - INFO - Loading 1 new model 2024-10-15 19:06:45,020 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 19:06:46,571 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 19:06:46,572 - root - INFO - Loading 1 new model 2024-10-15 19:06:46,603 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 19:06:46,809 - root - INFO - Prompt executed in 4.13 seconds 2024-10-15 19:06:51,229 - root - INFO - got prompt 2024-10-15 19:06:53,020 - root - INFO - Prompt executed in 1.78 seconds 2024-10-15 19:06:57,057 - root - INFO - got prompt 2024-10-15 19:06:58,873 - root - INFO - Prompt executed in 1.81 seconds 2024-10-15 19:10:06,384 - root - INFO - got prompt 2024-10-15 19:10:08,158 - root - INFO - Prompt executed in 1.76 seconds 2024-10-15 20:09:11,769 - root - INFO - got prompt 2024-10-15 20:09:13,548 - root - INFO - Prompt executed in 1.76 seconds 2024-10-15 20:09:27,885 - root - INFO - got prompt 2024-10-15 20:09:29,650 - root - INFO - Prompt executed in 1.75 seconds 2024-10-15 20:09:53,947 - root - INFO - got prompt 2024-10-15 20:09:55,727 - root - INFO - Prompt executed in 1.77 seconds 2024-10-15 20:10:18,289 - root - INFO - got prompt 2024-10-15 20:10:21,200 - root - INFO - Prompt executed in 2.90 seconds 2024-10-15 20:10:26,218 - root - INFO - got prompt 2024-10-15 20:10:29,144 - root - INFO - Prompt executed in 2.92 seconds 2024-10-15 20:10:40,864 - root - INFO - got prompt 2024-10-15 20:10:43,818 - root - INFO - Prompt executed in 2.94 seconds 2024-10-15 20:10:53,759 - root - INFO - got prompt 2024-10-15 20:10:56,703 - root - INFO - Prompt executed in 2.93 seconds 2024-10-15 20:11:33,009 - root - INFO - got prompt 2024-10-15 20:11:35,945 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 20:11:37,763 - root - INFO - got prompt 2024-10-15 20:11:40,699 - root - INFO - Prompt executed in 2.93 seconds 2024-10-15 20:11:50,443 - root - INFO - got prompt 2024-10-15 20:11:50,457 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:11:50,457 - root - INFO - Loading 1 new model 2024-10-15 20:11:51,625 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:11:51,640 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 20:11:51,644 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 20:11:51,647 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 20:11:54,303 - root - INFO - Prompt executed in 3.85 seconds 2024-10-15 20:12:04,988 - root - INFO - got prompt 2024-10-15 20:12:05,007 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:12:05,008 - root - INFO - Loading 1 new model 2024-10-15 20:12:05,552 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:12:05,719 - root - INFO - Requested to load Flux 2024-10-15 20:12:05,719 - root - INFO - Loading 1 new model 2024-10-15 20:12:07,295 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 20:12:09,837 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 20:12:09,838 - root - INFO - Loading 1 new model 2024-10-15 20:12:09,866 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 20:12:10,234 - root - INFO - Prompt executed in 5.23 seconds 2024-10-15 20:12:17,756 - root - INFO - got prompt 2024-10-15 20:12:17,772 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:12:17,773 - root - INFO - Loading 1 new model 2024-10-15 20:12:18,230 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:12:18,244 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 20:12:18,248 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 20:12:18,251 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 20:12:20,278 - root - INFO - Prompt executed in 2.51 seconds 2024-10-15 20:19:49,394 - root - INFO - got prompt 2024-10-15 20:19:49,410 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:19:49,411 - root - INFO - Loading 1 new model 2024-10-15 20:19:49,862 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:19:50,031 - root - INFO - Requested to load Flux 2024-10-15 20:19:50,031 - root - INFO - Loading 1 new model 2024-10-15 20:19:51,786 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 20:19:54,308 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 20:19:54,308 - root - INFO - Loading 1 new model 2024-10-15 20:19:54,336 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 20:19:54,694 - root - INFO - Prompt executed in 5.29 seconds 2024-10-15 20:20:00,460 - root - INFO - got prompt 2024-10-15 20:20:03,370 - root - INFO - Prompt executed in 2.90 seconds 2024-10-15 20:20:17,158 - root - INFO - got prompt 2024-10-15 20:20:20,083 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 20:20:32,943 - root - INFO - got prompt 2024-10-15 20:20:39,714 - root - INFO - Prompt executed in 6.76 seconds 2024-10-15 20:20:49,453 - root - INFO - got prompt 2024-10-15 20:20:52,403 - root - INFO - Prompt executed in 2.94 seconds 2024-10-15 20:21:08,416 - root - INFO - got prompt 2024-10-15 20:21:11,347 - root - INFO - Prompt executed in 2.92 seconds 2024-10-15 20:21:31,028 - root - INFO - got prompt 2024-10-15 20:21:34,991 - root - INFO - Prompt executed in 3.95 seconds 2024-10-15 20:21:43,935 - root - INFO - got prompt 2024-10-15 20:21:46,853 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 20:21:54,293 - root - INFO - got prompt 2024-10-15 20:21:57,185 - root - INFO - Prompt executed in 2.88 seconds 2024-10-15 20:22:25,668 - root - INFO - got prompt 2024-10-15 20:22:25,715 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:22:25,715 - root - INFO - Loading 1 new model 2024-10-15 20:22:26,833 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:22:26,848 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 20:22:26,852 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 20:22:26,854 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 20:22:29,477 - root - INFO - Prompt executed in 3.80 seconds 2024-10-15 20:22:32,171 - root - INFO - got prompt 2024-10-15 20:22:32,198 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:22:32,199 - root - INFO - Loading 1 new model 2024-10-15 20:22:32,636 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:22:32,940 - root - INFO - Requested to load Flux 2024-10-15 20:22:32,940 - root - INFO - Loading 1 new model 2024-10-15 20:22:34,761 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 20:22:37,300 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 20:22:37,300 - root - INFO - Loading 1 new model 2024-10-15 20:22:37,327 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 20:22:37,680 - root - INFO - Prompt executed in 5.50 seconds 2024-10-15 20:23:14,349 - root - INFO - got prompt 2024-10-15 20:23:17,251 - root - INFO - Prompt executed in 2.89 seconds 2024-10-15 20:23:22,653 - root - INFO - got prompt 2024-10-15 20:23:25,577 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 20:23:31,967 - root - INFO - got prompt 2024-10-15 20:23:31,992 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:23:31,993 - root - INFO - Loading 1 new model 2024-10-15 20:23:32,768 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:23:32,783 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 20:23:32,786 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 20:23:32,788 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 20:23:34,910 - root - INFO - Prompt executed in 2.93 seconds 2024-10-15 20:24:27,861 - root - INFO - got prompt 2024-10-15 20:24:27,887 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:24:27,887 - root - INFO - Loading 1 new model 2024-10-15 20:24:28,319 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:24:28,619 - root - INFO - Requested to load Flux 2024-10-15 20:24:28,619 - root - INFO - Loading 1 new model 2024-10-15 20:24:30,358 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 20:24:32,888 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 20:24:32,888 - root - INFO - Loading 1 new model 2024-10-15 20:24:32,918 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 20:24:33,294 - root - INFO - Prompt executed in 5.42 seconds 2024-10-15 20:24:48,094 - root - INFO - got prompt 2024-10-15 20:24:51,032 - root - INFO - Prompt executed in 2.93 seconds 2024-10-15 20:25:24,535 - root - INFO - got prompt 2024-10-15 20:25:27,436 - root - INFO - Prompt executed in 2.89 seconds 2024-10-15 20:25:31,078 - root - INFO - got prompt 2024-10-15 20:25:33,995 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 20:25:54,309 - root - INFO - got prompt 2024-10-15 20:25:54,346 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:25:54,347 - root - INFO - Loading 1 new model 2024-10-15 20:25:54,791 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:25:54,807 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 20:25:54,812 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 20:25:54,815 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 20:25:57,231 - root - INFO - Prompt executed in 2.91 seconds 2024-10-15 20:26:00,630 - root - INFO - got prompt 2024-10-15 20:26:00,657 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:26:00,658 - root - INFO - Loading 1 new model 2024-10-15 20:26:01,061 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:26:01,365 - root - INFO - Requested to load Flux 2024-10-15 20:26:01,365 - root - INFO - Loading 1 new model 2024-10-15 20:26:03,047 - root - INFO - loaded completely 0.0 12099.961059570312 True 2024-10-15 20:26:05,592 - root - INFO - Requested to load AutoencodingEngine 2024-10-15 20:26:05,592 - root - INFO - Loading 1 new model 2024-10-15 20:26:05,623 - root - INFO - loaded completely 0.0 159.87335777282715 True 2024-10-15 20:26:05,989 - root - INFO - Prompt executed in 5.35 seconds 2024-10-15 20:26:35,732 - root - INFO - got prompt 2024-10-15 20:26:35,759 - root - INFO - Requested to load FluxClipModel 2024-10-15 20:26:35,759 - root - INFO - Loading 1 new model 2024-10-15 20:26:36,199 - root - INFO - loaded completely 0.0 5062.70263671875 True 2024-10-15 20:26:36,214 - root - ERROR - !!! Exception during processing !!! Allocation on device 2024-10-15 20:26:36,218 - root - ERROR - Traceback (most recent call last): File "/workspace/ComfyUI/execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "/workspace/ComfyUI/execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "/workspace/ComfyUI/execution.py", line 158, in process_inputs results.append(getattr(obj, func)(inputs)) File "/workspace/ComfyUI/nodes.py", line 65, in encode output = clip.encode_from_tokens(tokens, return_pooled=True, return_dict=True) File "/workspace/ComfyUI/comfy/sd.py", line 125, in encode_from_tokens o = self.cond_stage_model.encode_token_weights(tokens) File "/workspace/ComfyUI/comfy/text_encoders/flux.py", line 59, in encode_token_weights t5_out, t5_pooled = self.t5xxl.encode_token_weights(token_weight_pairs_t5) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 41, in encode_token_weights o = self.encode(to_encode) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 229, in encode return self(tokens) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(args, kwargs) File "/workspace/ComfyUI/comfy/sd1_clip.py", line 201, in forward outputs = self.transformer(tokens, attention_mask_model, intermediate_output=self.layer_idx, final_layer_norm_intermediate=self.layer_norm_hidden_state, dtype=torch.float32) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, *kwargs) File "/workspace/ComfyUI/comfy/text_encoders/t5.py", line 238, in forward x = self.shared(input_ids, out_dtype=kwargs.get("dtype", torch.float32)) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(args, kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, kwargs) File "/workspace/ComfyUI/comfy/ops.py", line 211, in forward return self.forward_comfy_cast_weights(*args, *kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 152, in forward_comfy_cast_weights return self.forward_ggml_cast_weights(input, args, kwargs) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 187, in forward_ggml_cast_weights weight, _bias = self.cast_bias_weight(self, device=input.device, dtype=out_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 146, in cast_bias_weight weight = s.get_weight(s.weight.to(device), dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 119, in get_weight weight = dequantize_tensor(tensor, dtype, self.dequant_dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/dequant.py", line 23, in dequantize_tensor return dequantize(tensor.data, qtype, oshape, dtype=dequant_dtype).to(dtype) File "/workspace/ComfyUI/custom_nodes/ComfyUI-GGUF/ops.py", line 23, in to new = super().to(*args, *kwargs) File "/opt/environments/python/comfyui/lib/python3.10/site-packages/torch/_tensor.py", line 1437, in __torch_function__ ret = func(args, **kwargs) torch.OutOfMemoryError: Allocation on device
2024-10-15 20:26:36,220 - root - ERROR - Got an OOM, unloading all loaded models. 2024-10-15 20:26:38,249 - root - INFO - Prompt executed in 2.51 seconds
## Attached Workflow
Please make sure that workflow does not contain any sensitive information such as API keys or passwords.
{"last_node_id":12,"last_link_id":13,"nodes":[{"id":12,"type":"VAELoader","pos":{"0":862,"1":504},"size":{"0":315,"1":58},"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"VAE","type":"VAE","links":[13],"shape":3}],"properties":{"Node name for S&R":"VAELoader"},"widgets_values":["ae.safetensors"]},{"id":8,"type":"VAEDecode","pos":{"0":904,"1":55},"size":{"0":210,"1":46},"flags":{},"order":7,"mode":0,"inputs":[{"name":"samples","type":"LATENT","link":7},{"name":"vae","type":"VAE","link":13}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[9],"slot_index":0}],"properties":{"Node name for S&R":"VAEDecode"}},{"id":7,"type":"CLIPTextEncode","pos":{"0":380,"1":514},"size":{"0":427.3594970703125,"1":76},"flags":{},"order":5,"mode":0,"inputs":[{"name":"clip","type":"CLIP","link":12}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[6],"slot_index":0}],"properties":{"Node name for S&R":"CLIPTextEncode"},"widgets_values":[""]},{"id":10,"type":"UnetLoaderGGUF","pos":{"0":480,"1":193},"size":{"0":315,"1":58},"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"MODEL","type":"MODEL","links":[10],"shape":3}],"properties":{"Node name for S&R":"UnetLoaderGGUF"},"widgets_values":["flux1-schnell-Q8_0.gguf"]},{"id":11,"type":"DualCLIPLoaderGGUF","pos":{"0":485,"1":38},"size":{"0":315,"1":106},"flags":{},"order":2,"mode":0,"inputs":[],"outputs":[{"name":"CLIP","type":"CLIP","links":[11,12],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"DualCLIPLoaderGGUF"},"widgets_values":["clip_l.safetensors","t5-v1_1-xxl-encoder-Q8_0.gguf","flux"]},{"id":9,"type":"SaveImage","pos":{"0":1202,"1":56},"size":[767.8958005578725,813.7210937499999],"flags":{},"order":8,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":9}],"outputs":[],"properties":{},"widgets_values":["ComfyUI"]},{"id":5,"type":"EmptyLatentImage","pos":{"0":494,"1":646},"size":{"0":315,"1":106},"flags":{},"order":3,"mode":0,"inputs":[],"outputs":[{"name":"LATENT","type":"LATENT","links":[2],"slot_index":0}],"properties":{"Node name for S&R":"EmptyLatentImage"},"widgets_values":[1024,1024,1]},{"id":3,"type":"KSampler","pos":{"0":863,"1":186},"size":{"0":315,"1":262},"flags":{},"order":6,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":10},{"name":"positive","type":"CONDITIONING","link":4},{"name":"negative","type":"CONDITIONING","link":6},{"name":"latent_image","type":"LATENT","link":2}],"outputs":[{"name":"LATENT","type":"LATENT","links":[7],"slot_index":0}],"properties":{"Node name for S&R":"KSampler"},"widgets_values":[224070116344688,"randomize",4,1,"dpmpp_2m","simple",1]},{"id":6,"type":"CLIPTextEncode","pos":{"0":400,"1":308},"size":{"0":421.8337707519531,"1":133.60507202148438},"flags":{},"order":4,"mode":0,"inputs":[{"name":"clip","type":"CLIP","link":11}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[4],"slot_index":0}],"properties":{"Node name for S&R":"CLIPTextEncode"},"widgets_values":["(masterpiece), best quality, expressive eyes, perfect face, , sam yang, (Joe Madureira:0.8), (Extremely detailed), (Intricate details), Beautiful girl, tall and slender fit body, large breasts, shiny pale oiled skin, (Very long lush shiny straight (((iridescent white))) hair), (Extremely detailed beautiful vivid pale grey eyes), (Perfect make-up, mascara, eyeliner, blush), (Very long very fluffy fake lashes), (Delicate face features), ((Very long sharp shiny black nails)), (Intricate massive jewelry), Very large earloops earrings, black brows, Shiny glossy lips, (Cowboy shot), (Shiny navel piercing), real, hyperrealisim, super realistic skin texture,"]}],"links":[[2,5,0,3,3,"LATENT"],[4,6,0,3,1,"CONDITIONING"],[6,7,0,3,2,"CONDITIONING"],[7,3,0,8,0,"LATENT"],[9,8,0,9,0,"IMAGE"],[10,10,0,3,0,"MODEL"],[11,11,0,6,0,"CLIP"],[12,11,0,7,0,"CLIP"],[13,12,0,8,1,"VAE"]],"groups":[],"config":{},"extra":{"ds":{"scale":1.01,"offset":[-346.0182823003465,4.376936704119686]}},"version":0.4}
Expected Behavior
workflow run normally
Actual Behavior
Hi, getting CLIPTextEncode Allocation on device after update comfyui.., with the same workflow ( didnt modify anything ) and before all was good.
Steps to Reproduce
use any workflow with the new version
Debug Logs
Other
No response