Closed ddysDDYS closed 1 week ago
controlnet not work
1 2 3 4
1
Microsoft Edge
2024-06-20 10:52:17,233 - ControlNet - INFO - ControlNet Method tile_resample patched. To load target model SDXL Begin to load 1 model [Memory Management] Current Free GPU Memory (MB) = 2871.4102544784546 [Memory Management] Model Memory (MB) = 4897.086494445801 [Memory Management] Minimal Inference Memory (MB) = 1024.0 [Memory Management] Estimated Remaining GPU Memory (MB) = -3049.676239967346 [Memory Management] Requested SYNC Preserved Memory (MB) = 1421.084810256958 [Memory Management] Parameters Loaded to SYNC Stream (MB) = 3475.980842590332 [Memory Management] Parameters Loaded to GPU (MB) = 1421.0675048828125 Moving model(s) has taken 14.42 seconds 0%| | 0/20 [00:00<?, ?it/s] Traceback (most recent call last): File "C:\Users\20389\Desktop\sdxl\webui\modules_forge\main_thread.py", line 37, in loop task.work() File "C:\Users\20389\Desktop\sdxl\webui\modules_forge\main_thread.py", line 26, in work self.result = self.func(*self.args, **self.kwargs) File "C:\Users\20389\Desktop\sdxl\webui\modules\txt2img.py", line 111, in txt2img_function processed = processing.process_images(p) File "C:\Users\20389\Desktop\sdxl\webui\modules\processing.py", line 752, in process_images res = process_images_inner(p) File "C:\Users\20389\Desktop\sdxl\webui\modules\processing.py", line 922, in process_images_inner samples_ddim = p.sample(conditioning=p.c, unconditional_conditioning=p.uc, seeds=p.seeds, subseeds=p.subseeds, subseed_strength=p.subseed_strength, prompts=p.prompts) File "C:\Users\20389\Desktop\sdxl\webui\modules\processing.py", line 1275, in sample samples = self.sampler.sample(self, x, conditioning, unconditional_conditioning, image_conditioning=self.txt2img_image_conditioning(x)) File "C:\Users\20389\Desktop\sdxl\webui\modules\sd_samplers_kdiffusion.py", line 251, in sample samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args=self.sampler_extra_args, disable=False, callback=self.callback_state, **extra_params_kwargs)) File "C:\Users\20389\Desktop\sdxl\webui\modules\sd_samplers_common.py", line 263, in launch_sampling return func() File "C:\Users\20389\Desktop\sdxl\webui\modules\sd_samplers_kdiffusion.py", line 251, in <lambda> samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args=self.sampler_extra_args, disable=False, callback=self.callback_state, **extra_params_kwargs)) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context return func(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\webui\repositories\k-diffusion\k_diffusion\sampling.py", line 145, in sample_euler_ancestral denoised = model(x, sigmas[i] * s_in, **extra_args) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\webui\modules\sd_samplers_cfg_denoiser.py", line 182, in forward denoised = forge_sampler.forge_sample(self, denoiser_params=denoiser_params, File "C:\Users\20389\Desktop\sdxl\webui\modules_forge\forge_sampler.py", line 88, in forge_sample denoised = sampling_function(model, x, timestep, uncond, cond, cond_scale, model_options, seed) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\modules\samplers.py", line 289, in sampling_function cond_pred, uncond_pred = calc_cond_uncond_batch(model, cond, uncond_, x, timestep, model_options) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\modules\samplers.py", line 258, in calc_cond_uncond_batch output = model.apply_model(input_x, timestep_, **c).chunk(batch_chunks) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\modules\model_base.py", line 90, in apply_model model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float() File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\ldm\modules\diffusionmodules\openaimodel.py", line 867, in forward h = forward_timestep_embed(module, h, emb, context, transformer_options, time_context=time_context, num_video_frames=num_video_frames, image_only_indicator=image_only_indicator) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\ldm\modules\diffusionmodules\openaimodel.py", line 55, in forward_timestep_embed x = layer(x, context, transformer_options) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\ldm\modules\attention.py", line 620, in forward x = block(x, context=context[i], transformer_options=transformer_options) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\ldm\modules\attention.py", line 447, in forward return checkpoint(self._forward, (x, context, transformer_options), self.parameters(), self.checkpoint) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\ldm\modules\diffusionmodules\util.py", line 194, in checkpoint return func(*inputs) File "C:\Users\20389\Desktop\sdxl\webui\ldm_patched\ldm\modules\attention.py", line 486, in _forward n, context_attn1, value_attn1 = p(n, context_attn1, value_attn1, extra_options) File "C:\Users\20389\Desktop\sdxl\webui\extensions-builtin\sd_forge_controlllite\lib_controllllite\lib_controllllite.py", line 102, in __call__ q = q + self.modules[module_pfx_to_q](q) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\webui\extensions-builtin\sd_forge_controlllite\lib_controllllite\lib_controllllite.py", line 216, in forward cx = self.conditioning1(self.cond_image.to(x.device, dtype=x.dtype)) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\container.py", line 215, in forward input = module(input) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\conv.py", line 460, in forward return self._conv_forward(input, self.weight, self.bias) File "C:\Users\20389\Desktop\sdxl\system\python\lib\site-packages\torch\nn\modules\conv.py", line 456, in _conv_forward return F.conv2d(input, weight, bias, self.stride, RuntimeError: Input type (float) and bias type (struct c10::Half) should be the same Input type (float) and bias type (struct c10::Half) should be the same *** Error completing request *** Arguments: ('task(pkrslkwbbx5sq3w)', <gradio.routes.Request object at 0x000001E0CD17A4D0>, 'score_9,score_8_up,score_7_up,rating_explicit BREAK (1futa have pussy and penis and anus:1.2),oil,double bun,wavy hair,red hair,aqua eyes,gigantic breasts,tank_top,skirt_hold,white_pantyhose,black_loafers,maid_headdress,smile,standing,<lora:Retro_pixel_shading_pony-000010:1>,', 'bad_feet,bad_hands,bad anatomy,error,(worst quality, low quality:1.1),multiple tails,ugly,distorted,monster,score_4,multiple tails,sweat,', [], 20, 'Euler a', 1, 1, 7, 1216, 832, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, 'MultiDiffusion', False, True, 1024, 1024, 96, 96, 48, 4, 'None', 2, False, 10, 1, 1, 64, False, False, False, False, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 'DemoFusion', False, 128, 64, 4, 2, False, 10, 1, 1, 64, False, True, 3, 1, 1, True, 0.85, 0.6, 4, False, False, 960, 64, True, True, True, False, False, 7, 100, 'Constant', 0, 'Constant', 0, 4, True, 'MEAN', 'AD', 1, 0, 1, ControlNetUnit(input_mode=<InputMode.SIMPLE: 'simple'>, use_preview_as_input=False, batch_image_dir='', batch_mask_dir='', batch_input_gallery=[], batch_mask_gallery=[], generated_image=None, mask_image=None, hr_option='Both', enabled=True, module='tile_resample', model='controlnetxlCNXL_bdsqlszTileAnime [0dbb6686]', weight=1, image={'image': array([[[ 16, 14, 15], *** [ 17, 16, 14], *** [ 20, 18, 19], *** ..., *** [ 65, 1, 35], *** [ 65, 1, 35], *** [ 78, 1, 43]], *** *** [[ 16, 14, 15], *** [ 24, 22, 23], *** [ 23, 21, 22], *** ..., *** [ 88, 1, 52], *** [ 74, 2, 39], *** [ 85, 1, 50]], *** *** [[ 25, 23, 24], *** [ 22, 20, 21], *** [ 15, 13, 14], *** ..., *** [ 89, 2, 53], *** [ 83, 1, 49], *** [ 73, 1, 39]], *** *** ..., *** *** [[ 86, 9, 53], *** [113, 13, 75], *** [ 98, 10, 60], *** ..., *** [ 79, 39, 47], *** [ 97, 54, 63], *** [ 94, 53, 59]], *** *** [[ 86, 9, 53], *** [104, 11, 64], *** [ 86, 10, 49], *** ..., *** [ 99, 56, 63], *** [100, 57, 64], *** [101, 58, 67]], *** *** [[ 89, 10, 55], *** [112, 12, 74], *** [116, 14, 80], *** ..., *** [106, 60, 70], *** [ 93, 52, 56], *** [ 94, 53, 61]]], dtype=uint8), 'mask': array([[[0, 0, 0], *** [0, 0, 0], *** [0, 0, 0], *** ..., *** [0, 0, 0], *** [0, 0, 0], *** [0, 0, 0]], *** *** [[0, 0, 0], *** [0, 0, 0], *** [0, 0, 0], *** ..., *** [0, 0, 0], *** [0, 0, 0], *** [0, 0, 0]], *** *** [[0, 0, 0], *** [0, 0, 0], *** [0, 0, 0], *** ..., *** [0, 0, 0], *** [0, 0, 0], *** [0, 0, 0]], *** *** ..., *** *** [[0, 0, 0], *** [0, 0, 0], *** [0, 0, 0], *** ..., *** [0, 0, 0], *** [0, 0, 0], *** [0, 0, 0]], *** *** [[0, 0, 0], *** [0, 0, 0], *** [0, 0, 0], *** ..., *** [0, 0, 0], *** [0, 0, 0], *** [0, 0, 0]], *** *** [[0, 0, 0], *** [0, 0, 0], *** [0, 0, 0], *** ..., *** [0, 0, 0], *** [0, 0, 0], *** [0, 0, 0]]], dtype=uint8)}, resize_mode='Crop and Resize', processor_res=0.5, threshold_a=0.5, threshold_b=0.5, guidance_start=0, guidance_end=1, pixel_perfect=False, control_mode='Balanced', save_detected_map=True), ControlNetUnit(input_mode=<InputMode.SIMPLE: 'simple'>, use_preview_as_input=False, batch_image_dir='', batch_mask_dir='', batch_input_gallery=[], batch_mask_gallery=[], generated_image=None, mask_image=None, hr_option='Both', enabled=False, module='None', model='None', weight=1, image=None, resize_mode='Crop and Resize', processor_res=-1, threshold_a=-1, threshold_b=-1, guidance_start=0, guidance_end=1, pixel_perfect=False, control_mode='Balanced', save_detected_map=True), ControlNetUnit(input_mode=<InputMode.SIMPLE: 'simple'>, use_preview_as_input=False, batch_image_dir='', batch_mask_dir='', batch_input_gallery=[], batch_mask_gallery=[], generated_image=None, mask_image=None, hr_option='Both', enabled=False, module='None', model='None', weight=1, image=None, resize_mode='Crop and Resize', processor_res=-1, threshold_a=-1, threshold_b=-1, guidance_start=0, guidance_end=1, pixel_perfect=False, control_mode='Balanced', save_detected_map=True), False, 7, 1, 'Constant', 0, 'Constant', 0, 1, 'enable', 'MEAN', 'AD', 1, False, 1.01, 1.02, 0.99, 0.95, False, 0.5, 2, False, 256, 2, 0, False, True, 3, 2, 0, 0.35, True, 'bicubic', 'bicubic', False, 0, 'anisotropic', 0, 'reinhard', 100, 0, 'subtract', 0, 0, 'gaussian', 'add', 0, 100, 127, 0, 'hard_clamp', 5, 0, 'None', 'None', False, 'MultiDiffusion', 768, 768, 64, 4, False, False, False, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {} Traceback (most recent call last): File "C:\Users\20389\Desktop\sdxl\webui\modules\call_queue.py", line 57, in f res = list(func(*args, **kwargs)) TypeError: 'NoneType' object is not iterable
No response
Checklist
What happened?
controlnet not work
Steps to reproduce the problem
1 2 3 4
What should have happened?
1
What browsers do you use to access the UI ?
Microsoft Edge
Sysinfo
Console logs
Additional information
No response