AUTOMATIC1111 / stable-diffusion-webui

Stable Diffusion web UI
GNU Affero General Public License v3.0
139.61k stars 26.47k forks source link

[Bug]: Sizes of tensors must match except in dimension 0 #14607

Open elfgoose opened 8 months ago

elfgoose commented 8 months ago

Checklist

What happened?

I'm having a strange issue where sometimes an instance will throw up a "size of tensors" error. Often it works fine, then subsequent generations, with no changes, throw up the error. restarting the CMD window usually (but not always) fixes it, but then it can happen again. If I have several browser windows open with SD, some might give the error, but others do not. I tried changing the model, in case the one I was using was corrupted or something, but I got the same issue

Steps to reproduce the problem

intermittent, but happens when I click generate

What should have happened?

it should generate without error

What browsers do you use to access the UI ?

Mozilla Firefox

Sysinfo

{ "Platform": "Windows-10-10.0.19045-SP0", "Python": "3.10.6", "Version": "v1.7.0", "Commit": "cf2772fab0af5573da775e7437e6acdca424f26e", "Script path": "C:\AI\Image AI\SDNEW\webui", "Data path": "C:\AI\Image AI\SDNEW\webui", "Extensions dir": "C:\AI\Image AI\SDNEW\webui\extensions", "Checksum": "4e6a90e49e2ce7e352f5453ee967299ff6b3dacc9e7344db71f371bdf499f0bd", "Commandline": [ "launch.py", "--listen", "--enable-insecure-extension-access", "--reinstall-xformers", "--xformers", "--no-half", "--no-half-vae", "--medvram", "--medvram-sdxl", "--update-all-extensions" ], "Torch env info": "'NoneType' object has no attribute 'splitlines'", "Exceptions": [ { "exception": "Sizes of tensors must match except in dimension 0. Expected size 1280 but got size 768 for tensor number 1 in the list.", "traceback": [ [ "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py, line 57, f", "res = list(func(*args, kwargs))" ], [ "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py, line 36, f", "res = func(*args, *kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\txt2img.py, line 55, txt2img", "processed = processing.process_images(p)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 734, process_images", "res = process_images_inner(p)" ], [ "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-controlnet\scripts\batch_hijack.py, line 42, processing_process_images_hijack", "return getattr(processing, '__controlnet_original_process_images_inner')(p, args, kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 857, process_images_inner", "p.setup_conds()" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 1308, setup_conds", "super().setup_conds()" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 470, setup_conds", "self.c = self.get_conds_with_caching(prompt_parser.get_multicond_learned_conditioning, prompts, total_steps, [self.cached_c], self.extra_network_data)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 455, get_conds_with_caching", "cache[1] = function(shared.sd_model, required_prompts, steps, hires_steps, shared.opts.use_old_scheduling)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\prompt_parser.py, line 261, get_multicond_learned_conditioning", "learned_conditioning = get_learned_conditioning(model, prompt_flat_list, steps, hires_steps, use_old_scheduling)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\prompt_parser.py, line 188, get_learned_conditioning", "conds = model.get_learned_conditioning(texts)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_models_xl.py, line 31, get_learned_conditioning", "c = self.conditioner(sdxl_conds, force_zero_embeddings=['txt'] if force_zero_negative_prompt else [])" ], [ "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py, line 1538, _call_impl", "result = forward_call(*args, kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\repositories\generative-models\sgm\modules\encoders\modules.py, line 141, forward", "emb_out = embedder(batch[embedder.input_key])" ], [ "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py, line 1501, _call_impl", "return forward_call(*args, *kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_clip.py, line 234, forward", "z = self.process_tokens(tokens, multipliers)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_clip.py, line 273, process_tokens", "z = self.encode_with_transformers(tokens)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_open_clip.py, line 57, encode_with_transformers", "d = self.wrapped.encode_with_transformer(tokens)" ], [ "C:\AI\Image AI\SDNEW\webui\repositories\generative-models\sgm\modules\encoders\modules.py, line 467, encode_with_transformer", "x = self.model.token_embedding(text) # [batch_size, n_ctx, d_model]" ], [ "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py, line 1501, _call_impl", "return forward_call(args, kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack.py, line 359, forward", "tensor = torch.cat([tensor[0:offset + 1], emb[0:emb_len], tensor[offset + 1 + emb_len:]])" ] ] }, { "exception": "Sizes of tensors must match except in dimension 0. Expected size 1280 but got size 768 for tensor number 1 in the list.", "traceback": [ [ "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py, line 57, f", "res = list(func(*args, kwargs))" ], [ "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py, line 36, f", "res = func(*args, *kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\txt2img.py, line 55, txt2img", "processed = processing.process_images(p)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 734, process_images", "res = process_images_inner(p)" ], [ "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-controlnet\scripts\batch_hijack.py, line 42, processing_process_images_hijack", "return getattr(processing, '__controlnet_original_process_images_inner')(p, args, kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 857, process_images_inner", "p.setup_conds()" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 1308, setup_conds", "super().setup_conds()" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 470, setup_conds", "self.c = self.get_conds_with_caching(prompt_parser.get_multicond_learned_conditioning, prompts, total_steps, [self.cached_c], self.extra_network_data)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\processing.py, line 455, get_conds_with_caching", "cache[1] = function(shared.sd_model, required_prompts, steps, hires_steps, shared.opts.use_old_scheduling)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\prompt_parser.py, line 261, get_multicond_learned_conditioning", "learned_conditioning = get_learned_conditioning(model, prompt_flat_list, steps, hires_steps, use_old_scheduling)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\prompt_parser.py, line 188, get_learned_conditioning", "conds = model.get_learned_conditioning(texts)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_models_xl.py, line 31, get_learned_conditioning", "c = self.conditioner(sdxl_conds, force_zero_embeddings=['txt'] if force_zero_negative_prompt else [])" ], [ "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py, line 1538, _call_impl", "result = forward_call(*args, kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\repositories\generative-models\sgm\modules\encoders\modules.py, line 141, forward", "emb_out = embedder(batch[embedder.input_key])" ], [ "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py, line 1501, _call_impl", "return forward_call(*args, *kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_clip.py, line 234, forward", "z = self.process_tokens(tokens, multipliers)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_clip.py, line 273, process_tokens", "z = self.encode_with_transformers(tokens)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_open_clip.py, line 57, encode_with_transformers", "d = self.wrapped.encode_with_transformer(tokens)" ], [ "C:\AI\Image AI\SDNEW\webui\repositories\generative-models\sgm\modules\encoders\modules.py, line 467, encode_with_transformer", "x = self.model.token_embedding(text) # [batch_size, n_ctx, d_model]" ], [ "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py, line 1501, _call_impl", "return forward_call(args, kwargs)" ], [ "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack.py, line 359, forward", "tensor = torch.cat([tensor[0:offset + 1], emb[0:emb_len], tensor[offset + 1 + emb_len:]])" ] ] }, { "exception": "Expecting value: line 1 column 1 (char 0)", "traceback": [ [ "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py, line 57, f", "res = list(func(*args, *kwargs))" ], [ "C:\AI\Image AI\SDNEW\webui\modules\ui_common.py, line 50, save_files", "data = json.loads(js_data)" ], [ "json\init.py, line 346, loads", "" ], [ "json\decoder.py, line 337, decode", "" ], [ "json\decoder.py, line 355, raw_decode", "" ] ] }, { "exception": "Expecting value: line 1 column 1 (char 0)", "traceback": [ [ "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py, line 57, f", "res = list(func(args, **kwargs))" ], [ "C:\AI\Image AI\SDNEW\webui\modules\ui_common.py, line 50, save_files", "data = json.loads(js_data)" ], [ "json\init.py, line 346, loads", "" ], [ "json\decoder.py, line 337, decode", "" ], [ "json\decoder.py, line 355, raw_decode", "" ] ] } ], "CPU": { "model": "Intel64 Family 6 Model 94 Stepping 3, GenuineIntel", "count logical": 8, "count physical": 4 }, "RAM": { "total": "32GB", "used": "26GB", "free": "6GB" }, "Extensions": [ { "name": "a1111-sd-webui-tagcomplete", "path": "C:\AI\Image AI\SDNEW\webui\extensions\a1111-sd-webui-tagcomplete", "version": "231b121f", "branch": "main", "remote": "https://github.com/DominikDoom/a1111-sd-webui-tagcomplete.git" }, { "name": "adetailer", "path": "C:\AI\Image AI\SDNEW\webui\extensions\adetailer", "version": "92692191", "branch": "main", "remote": "https://github.com/Bing-su/adetailer.git" }, { "name": "clip-interrogator-ext", "path": "C:\AI\Image AI\SDNEW\webui\extensions\clip-interrogator-ext", "version": "0f1a4591", "branch": "main", "remote": "https://github.com/pharmapsychotic/clip-interrogator-ext.git" }, { "name": "sd-webui-aspect-ratio-helper", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-aspect-ratio-helper", "version": "99fcf9b0", "branch": "main", "remote": "https://github.com/thomasasfk/sd-webui-aspect-ratio-helper.git" }, { "name": "sd-webui-controlnet", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-controlnet", "version": "a52e607b", "branch": "main", "remote": "https://github.com/Mikubill/sd-webui-controlnet.git" }, { "name": "sd-webui-deepdanbooru-object-recognition", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-deepdanbooru-object-recognition", "version": "5b3e7cbf", "branch": "main", "remote": "https://github.com/Jibaku789/sd-webui-deepdanbooru-object-recognition.git" }, { "name": "sd-webui-prompt-all-in-one", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-prompt-all-in-one", "version": "658ae822", "branch": "main", "remote": "https://github.com/Physton/sd-webui-prompt-all-in-one" }, { "name": "sd-webui-reactor", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-reactor", "version": "4b1cfc70", "branch": "main", "remote": "https://github.com/Gourieff/sd-webui-reactor" }, { "name": "sd-webui-segment-anything", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-segment-anything", "version": "d80220ec", "branch": "master", "remote": "https://github.com/continue-revolution/sd-webui-segment-anything.git" }, { "name": "sdwebui-close-confirmation-dialogue", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sdwebui-close-confirmation-dialogue", "version": "f7cfa87d", "branch": "main", "remote": "https://github.com/w-e-w/sdwebui-close-confirmation-dialogue.git" }, { "name": "stable-diffusion-webui-model-toolkit", "path": "C:\AI\Image AI\SDNEW\webui\extensions\stable-diffusion-webui-model-toolkit", "version": "cf824587", "branch": "master", "remote": "https://github.com/arenasys/stable-diffusion-webui-model-toolkit.git" } ], "Inactive extensions": [ { "name": "sd-webui-animatediff", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-animatediff", "version": "55ecff36", "branch": "master", "remote": "https://github.com/continue-revolution/sd-webui-animatediff.git" }, { "name": "sd_smartprocess-subfolder-upscale", "path": "C:\AI\Image AI\SDNEW\webui\extensions\sd_smartprocess-subfolder-upscale", "version": "5750d94f", "branch": "main", "remote": "https://github.com/baejun10/sd_smartprocess-subfolder-upscale" } ], "Environment": { "COMMANDLINE_ARGS": " --listen --enable-insecure-extension-access --reinstall-xformers --xformers --no-half --no-half-vae --medvram --medvram-sdxl --update-all-extensions ", "GRADIO_ANALYTICS_ENABLED": "False", "XFORMERS_PACKAGE": "xformers==0.0.21" }, "Config": { "samples_save": true, "samples_format": "jpg", "samples_filename_pattern": "", "save_images_add_number": true, "grid_save": false, "grid_format": "png", "grid_extended_filename": false, "grid_only_if_multiple": true, "grid_prevent_empty_spots": false, "grid_zip_filename_pattern": "", "n_rows": -1, "font": "", "grid_text_active_color": "#000000", "grid_text_inactive_color": "#999999", "grid_background_color": "#ffffff", "enable_pnginfo": true, "save_txt": true, "save_images_before_face_restoration": false, "save_images_before_highres_fix": true, "save_images_before_color_correction": false, "save_mask": false, "save_mask_composite": false, "jpeg_quality": 80, "webp_lossless": false, "export_for_4chan": true, "img_downscale_threshold": 4.0, "target_side_length": 4000, "img_max_size_mp": 200, "use_original_name_batch": true, "use_upscaler_name_as_suffix": false, "save_selected_only": true, "save_init_img": false, "temp_dir": "", "clean_temp_dir_at_start": false, "save_incomplete_images": false, "outdir_samples": "P:\imgsrc\SD images", "outdir_txt2img_samples": "outputs/txt2img-images", "outdir_img2img_samples": "outputs/img2img-images", "outdir_extras_samples": "outputs/extras-images", "outdir_grids": "", "outdir_txt2img_grids": "outputs/txt2img-grids", "outdir_img2img_grids": "outputs/img2img-grids", "outdir_save": "log/images", "outdir_init_images": "outputs/init-images", "save_to_dirs": true, "grid_save_to_dirs": true, "use_save_to_dirs_for_ui": false, "directories_filename_pattern": "[date]", "directories_max_prompt_words": 11, "ESRGAN_tile": 192, "ESRGAN_tile_overlap": 11, "realesrgan_enabled_models": [ "R-ESRGAN 4x+", "R-ESRGAN 4x+ Anime6B" ], "upscaler_for_img2img": "4x-UltraSharp", "face_restoration": false, "face_restoration_model": "CodeFormer", "code_former_weight": 0.5, "face_restoration_unload": false, "auto_launch_browser": "Disable", "show_warnings": false, "show_gradio_deprecation_warnings": true, "memmon_poll_rate": 8, "samples_log_stdout": false, "multiple_tqdm": true, "print_hypernet_extra": false, "list_hidden_files": true, "disable_mmap_load_safetensors": false, "hide_ldm_prints": true, "api_enable_requests": true, "api_forbid_local_requests": true, "api_useragent": "", "unload_models_when_training": false, "pin_memory": false, "save_optimizer_state": false, "save_training_settings_to_txt": true, "dataset_filename_word_regex": "", "dataset_filename_join_string": " ", "training_image_repeats_per_epoch": 1, "training_write_csv_every": 500, "training_xattention_optimizations": false, "training_enable_tensorboard": false, "training_tensorboard_save_images": false, "training_tensorboard_flush_every": 120, "sd_model_checkpoint": "SDXL_1.0\OpenDalleV1.1.safetensors [673887ace8]", "sd_checkpoints_limit": 2, "sd_checkpoints_keep_in_cpu": true, "sd_checkpoint_cache": 0, "sd_unet": "Automatic", "enable_quantization": false, "enable_emphasis": true, "enable_batch_seeds": true, "comma_padding_backtrack": 20, "CLIP_stop_at_last_layers": 2, "upcast_attn": true, "randn_source": "GPU", "tiling": false, "hires_fix_refiner_pass": "second pass", "sdxl_crop_top": 0, "sdxl_crop_left": 0, "sdxl_refiner_low_aesthetic_score": 2.5, "sdxl_refiner_high_aesthetic_score": 6.0, "sd_vae_explanation": "VAE is a neural network that transforms a standard RGB\nimage into latent space representation and back. Latent space representation is what stable diffusion is working on during sampling\n(i.e. when the progress bar is between empty and full). For txt2img, VAE is used to create a resulting image after the sampling is finished.\nFor img2img, VAE is used to process user's input image before the sampling, and to create an image after sampling.", "sd_vae_checkpoint_cache": 0, "sd_vae": "None", "sd_vae_overrides_per_model_preferences": true, "auto_vae_precision": true, "sd_vae_encode_method": "Full", "sd_vae_decode_method": "Full", "inpainting_mask_weight": 1.0, "initial_noise_multiplier": 1.0, "img2img_extra_noise": 0.0, "img2img_color_correction": true, "img2img_fix_steps": false, "img2img_background_color": "#ffffff", "img2img_editor_height": 720, "img2img_sketch_default_brush_color": "#ffffff", "img2img_inpaint_mask_brush_color": "#ffffff", "img2img_inpaint_sketch_default_brush_color": "#ffffff", "return_mask": false, "return_mask_composite": false, "cross_attention_optimization": "Automatic", "s_min_uncond": 0.0, "token_merging_ratio": 0.0, "token_merging_ratio_img2img": 0.0, "token_merging_ratio_hr": 0.0, "pad_cond_uncond": true, "persistent_cond_cache": true, "batch_cond_uncond": true, "use_old_emphasis_implementation": false, "use_old_karras_scheduler_sigmas": false, "no_dpmpp_sde_batch_determinism": false, "use_old_hires_fix_width_height": false, "dont_fix_second_order_samplers_schedule": false, "hires_fix_use_firstpass_conds": false, "use_old_scheduling": false, "interrogate_keep_models_in_memory": false, "interrogate_return_ranks": false, "interrogate_clip_num_beams": 1, "interrogate_clip_min_length": 24, "interrogate_clip_max_length": 180, "interrogate_clip_dict_limit": 1500, "interrogate_clip_skip_categories": [], "interrogate_deepbooru_score_threshold": 0.5, "deepbooru_sort_alpha": true, "deepbooru_use_spaces": true, "deepbooru_escape": true, "deepbooru_filter_tags": "", "extra_networks_show_hidden_directories": true, "extra_networks_hidden_models": "When searched", "extra_networks_default_multiplier": 1.0, "extra_networks_card_width": 0, "extra_networks_card_height": 0, "extra_networks_card_text_scale": 1.0, "extra_networks_card_show_desc": true, "extra_networks_add_text_separator": " ", "ui_extra_networks_tab_reorder": "", "textual_inversion_print_at_load": false, "textual_inversion_add_hashes_to_infotext": true, "sd_hypernetwork": "None", "localization": "None", "gradio_theme": "Default", "gradio_themes_cache": true, "gallery_height": "", "return_grid": true, "do_not_show_images": false, "send_seed": true, "send_size": true, "js_modal_lightbox": true, "js_modal_lightbox_initially_zoomed": true, "js_modal_lightbox_gamepad": false, "js_modal_lightbox_gamepad_repeat": 250, "show_progress_in_title": true, "samplers_in_dropdown": true, "dimensions_and_batch_together": true, "keyedit_precision_attention": 0.1, "keyedit_precision_extra": 0.05, "keyedit_delimiters": ".,\/!?%^*;:{}=`~()", "keyedit_move": true, "quicksettings_list": [ "sd_model_checkpoint", "CLIP_stop_at_last_layers", "sd_vae", "face_restoration", "face_restoration_model" ], "ui_tab_order": [], "hidden_tabs": [], "ui_reorder_list": [], "hires_fix_show_sampler": true, "hires_fix_show_prompts": false, "disable_token_counters": false, "add_model_hash_to_info": true, "add_model_name_to_info": true, "add_user_name_to_info": false, "add_version_to_infotext": true, "disable_weights_auto_swap": true, "infotext_styles": "Apply if any", "show_progressbar": true, "live_previews_enable": true, "live_previews_image_format": "jpeg", "show_progress_grid": true, "show_progress_every_n_steps": 5, "show_progress_type": "Approx NN", "live_preview_allow_lowvram_full": false, "live_preview_content": "Prompt", "live_preview_refresh_period": 1000, "live_preview_fast_interrupt": false, "hide_samplers": [], "eta_ddim": 0.0, "eta_ancestral": 1.0, "ddim_discretize": "uniform", "s_churn": 0.0, "s_tmin": 0.0, "s_tmax": 0.0, "s_noise": 1.0, "k_sched_type": "Automatic", "sigma_min": 0.0, "sigma_max": 0.0, "rho": 0.0, "eta_noise_seed_delta": 0, "always_discard_next_to_last_sigma": false, "sgm_noise_multiplier": false, "uni_pc_variant": "bh1", "uni_pc_skip_type": "time_uniform", "uni_pc_order": 3, "uni_pc_lower_order_final": true, "postprocessing_enable_in_main_ui": [], "postprocessing_operation_order": [], "upscaling_max_images_in_cache": 5, "disabled_extensions": [ "sd-webui-animatediff", "sd_smartprocess-subfolder-upscale" ], "disable_all_extensions": "none", "restore_config_state_file": "", "sd_checkpoint_hash": "673887ace8a5c56388a10e8d576d08f609393d981ea4938511179b50f09bdcc4", "ldsr_steps": 100, "ldsr_cached": false, "SCUNET_tile": 256, "SCUNET_tile_overlap": 8, "SWIN_tile": 192, "SWIN_tile_overlap": 8, "lora_functional": false, "sd_lora": "None", "lora_preferred_name": "Alias from file", "lora_add_hashes_to_infotext": true, "lora_show_all": false, "lora_hide_unknown_for_versions": [], "lora_in_memory_limit": 0, "extra_options_txt2img": [ "face_restoration", "face_restoration_model" ], "extra_options_img2img": [ "face_restoration" ], "extra_options_cols": 1, "extra_options_accordion": false, "canvas_hotkey_zoom": "Alt", "canvas_hotkey_adjust": "Ctrl", "canvas_hotkey_move": "F", "canvas_hotkey_fullscreen": "S", "canvas_hotkey_reset": "R", "canvas_hotkey_overlap": "O", "canvas_show_tooltip": true, "canvas_auto_expand": true, "canvas_blur_prompt": false, "canvas_disabled_functions": [ "Overlap" ], "ad_max_models": 2, "ad_save_previews": false, "ad_save_images_before": false, "ad_only_seleted_scripts": true, "ad_script_names": "dynamic_prompting,dynamic_thresholding,wildcard_recursive,wildcards,lora_block_weight", "ad_bbox_sortby": "None", "control_net_detectedmap_dir": "detected_maps", "control_net_models_path": "", "control_net_modules_path": "", "control_net_unit_count": 3, "control_net_model_cache_size": 1, "control_net_inpaint_blur_sigma": 7, "control_net_no_high_res_fix": false, "control_net_no_detectmap": false, "control_net_detectmap_autosaving": false, "control_net_allow_script_control": false, "control_net_sync_field_args": true, "controlnet_show_batch_images_in_ui": false, "controlnet_increment_seed_during_batch": false, "controlnet_disable_control_type": false, "controlnet_disable_openpose_edit": false, "controlnet_ignore_noninpaint_mask": false, "tac_tagFile": "danbooru.csv", "tac_active": true, "tac_activeIn.txt2img": true, "tac_activeIn.img2img": true, "tac_activeIn.negativePrompts": true, "tac_activeIn.thirdParty": true, "tac_activeIn.modelList": "", "tac_activeIn.modelListMode": "Blacklist", "tac_slidingPopup": true, "tac_maxResults": 5.0, "tac_showAllResults": false, "tac_resultStepLength": 100.0, "tac_delayTime": 100.0, "tac_useWildcards": true, "tac_sortWildcardResults": true, "tac_useEmbeddings": true, "tac_includeEmbeddingsInNormalResults": false, "tac_useHypernetworks": true, "tac_useLoras": true, "tac_useLycos": true, "tac_showWikiLinks": false, "tac_showExtraNetworkPreviews": true, "tac_modelSortOrder": "Name", "tac_replaceUnderscores": true, "tac_escapeParentheses": true, "tac_appendComma": true, "tac_appendSpace": true, "tac_alwaysSpaceAtEnd": true, "tac_modelKeywordCompletion": "Never", "tac_modelKeywordLocation": "Start of prompt", "tac_wildcardCompletionMode": "To next folder level", "tac_alias.searchByAlias": true, "tac_alias.onlyShowAlias": false, "tac_translation.translationFile": "None", "tac_translation.oldFormat": false, "tac_translation.searchByTranslation": true, "tac_translation.liveTranslation": false, "tac_extra.extraFile": "extra-quality-tags.csv", "tac_extra.addMode": "Insert before", "tac_chantFile": "demo-chants.json", "tac_keymap": "{\n \"MoveUp\": \"ArrowUp\",\n \"MoveDown\": \"ArrowDown\",\n \"JumpUp\": \"PageUp\",\n \"JumpDown\": \"PageDown\",\n \"JumpToStart\": \"Home\",\n \"JumpToEnd\": \"End\",\n \"ChooseSelected\": \"Enter\",\n \"ChooseFirstOrSelected\": \"Tab\",\n \"Close\": \"Escape\"\n}", "tac_colormap": "{\n \"danbooru\": {\n \"-1\": [\"red\", \"maroon\"],\n \"0\": [\"lightblue\", \"dodgerblue\"],\n \"1\": [\"indianred\", \"firebrick\"],\n \"3\": [\"violet\", \"darkorchid\"],\n \"4\": [\"lightgreen\", \"darkgreen\"],\n \"5\": [\"orange\", \"darkorange\"]\n },\n \"e621\": {\n \"-1\": [\"red\", \"maroon\"],\n \"0\": [\"lightblue\", \"dodgerblue\"],\n \"1\": [\"gold\", \"goldenrod\"],\n \"3\": [\"violet\", \"darkorchid\"],\n \"4\": [\"lightgreen\", \"darkgreen\"],\n \"5\": [\"tomato\", \"darksalmon\"],\n \"6\": [\"red\", \"maroon\"],\n \"7\": [\"whitesmoke\", \"black\"],\n \"8\": [\"seagreen\", \"darkseagreen\"]\n }\n}", "tac_refreshTempFiles": "Refresh TAC temp files", "animatediff_model_path": "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-animatediff\model", "animatediff_optimize_gif_palette": true, "animatediff_optimize_gif_gifsicle": true, "animatediff_xformers": "Optimize attention layers with sdp (torch >= 2.0.0 required)", "arh_javascript_aspect_ratio_show": true, "arh_javascript_aspect_ratio": "1:1, 3:2, 4:3, 5:4, 16:9", "arh_ui_javascript_selection_method": "Aspect Ratios Dropdown", "arh_hide_accordion_by_default": true, "arh_expand_by_default": false, "arh_ui_component_order_key": "MaxDimensionScaler, MinDimensionScaler, PredefinedAspectRatioButtons, PredefinedPercentageButtons", "arh_show_max_width_or_height": false, "arh_max_width_or_height": 1024, "arh_show_min_width_or_height": false, "arh_min_width_or_height": 1024, "arh_show_predefined_aspect_ratios": false, "arh_predefined_aspect_ratio_use_max_dim": false, "arh_predefined_aspect_ratios": "1:1, 4:3, 16:9, 9:16, 21:9", "arh_show_predefined_percentages": false, "arh_predefined_percentages": "25, 50, 75, 125, 150, 175, 200", "arh_predefined_percentages_display_key": "Incremental/decremental percentage (-50%, +50%)", "sam_use_local_groundingdino": false, "ad_extra_models_dir": "", "ad_same_seed_for_each_tap": false, "hypertile_enable_unet": false, "hypertile_enable_unet_secondpass": false, "hypertile_max_depth_unet": 3, "hypertile_max_tile_unet": 256, "hypertile_swap_size_unet": 3, "hypertile_enable_vae": false, "hypertile_max_depth_vae": 3, "hypertile_max_tile_vae": 128, "hypertile_swap_size_vae": 3, "tac_wildcardExclusionList": "", "tac_skipWildcardRefresh": false, "model_toolkit_fix_clip": false, "model_toolkit_autoprune": false, "save_images_replace_action": "Replace", "notification_audio": true, "notification_volume": 100, "extra_networks_dir_button_function": false, "extra_networks_card_order_field": "Path", "extra_networks_card_order": "Ascending", "img2img_batch_show_results_limit": 32, "add_vae_name_to_info": true, "add_vae_hash_to_info": true, "infotext_skip_pasting": [], "js_live_preview_in_modal_lightbox": false, "keyedit_delimiters_whitespace": [ "Tab", "Carriage Return", "Line Feed" ], "compact_prompt_box": false, "sd_checkpoint_dropdown_use_short": false, "txt2img_settings_accordion": false, "img2img_settings_accordion": false, "enable_console_prompts": false, "dump_stacks_on_signal": false, "postprocessing_existing_caption_action": "Ignore", "tac_useLoraPrefixForLycos": true, "controlnet_disable_photopea_edit": false, "controlnet_photopea_warning": true }, "Startup": { "total": 113.66031813621521, "records": { "initial startup": 0.10100173950195312, "prepare environment/checks": 2.064997673034668, "prepare environment/git version info": 0.13800334930419922, "prepare environment/torch GPU test": 5.58609676361084, "prepare environment/install xformers": 7.65751051902771, "prepare environment/clone repositores": 10.569517850875854, "prepare environment/run extensions installers/a1111-sd-webui-tagcomplete": 0.00099945068359375, "prepare environment/run extensions installers/adetailer": 2.3112337589263916, "prepare environment/run extensions installers/clip-interrogator-ext": 18.06752586364746, "prepare environment/run extensions installers/sd-webui-aspect-ratio-helper": 0.0, "prepare environment/run extensions installers/sd-webui-controlnet": 3.5942256450653076, "prepare environment/run extensions installers/sd-webui-deepdanbooru-object-recognition": 0.0, "prepare environment/run extensions installers/sd-webui-prompt-all-in-one": 2.3106131553649902, "prepare environment/run extensions installers/sd-webui-reactor": 6.301990270614624, "prepare environment/run extensions installers/sd-webui-segment-anything": 2.2858686447143555, "prepare environment/run extensions installers/sdwebui-close-confirmation-dialogue": 0.0, "prepare environment/run extensions installers/stable-diffusion-webui-model-toolkit": 0.0, "prepare environment/run extensions installers": 34.872456789016724, "prepare environment/update extensions": 8.426710844039917, "prepare environment": 69.36929273605347, "launcher": 0.001999378204345703, "import torch": 15.4760000705719, "import gradio": 1.6190016269683838, "setup paths": 1.2499985694885254, "import ldm": 0.01100301742553711, "import sgm": 0.0, "initialize shared": 0.28999781608581543, "other imports": 1.2020020484924316, "opts onchange": 0.0, "setup SD model": 0.008000612258911133, "setup codeformer": 0.3619973659515381, "setup gfpgan": 0.04699993133544922, "set samplers": 0.0, "list extensions": 0.006001949310302734, "restore config state file": 0.0, "list SD models": 0.3739964962005615, "list localizations": 0.0010013580322265625, "load scripts/custom_code.py": 0.013998985290527344, "load scripts/img2imgalt.py": 0.0, "load scripts/loopback.py": 0.0010018348693847656, "load scripts/outpainting_mk_2.py": 0.0, "load scripts/poor_mans_outpainting.py": 0.0010004043579101562, "load scripts/postprocessing_caption.py": 0.0, "load scripts/postprocessing_codeformer.py": 0.00099945068359375, "load scripts/postprocessing_create_flipped_copies.py": 0.0, "load scripts/postprocessing_focal_crop.py": 0.0019979476928710938, "load scripts/postprocessing_gfpgan.py": 0.0, "load scripts/postprocessing_split_oversized.py": 0.0, "load scripts/postprocessing_upscale.py": 0.0010008811950683594, "load scripts/processing_autosized_crop.py": 0.0010004043579101562, "load scripts/prompt_matrix.py": 0.0020008087158203125, "load scripts/prompts_from_file.py": 0.001001119613647461, "load scripts/sd_upscale.py": 0.0009989738464355469, "load scripts/xyz_grid.py": 0.003999948501586914, "load scripts/ldsr_model.py": 1.552001714706421, "load scripts/lora_script.py": 1.1649980545043945, "load scripts/scunet_model.py": 0.13000059127807617, "load scripts/swinir_model.py": 0.2350003719329834, "load scripts/hotkey_config.py": 0.006999492645263672, "load scripts/extra_options_section.py": 0.0029993057250976562, "load scripts/hypertile_script.py": 0.27899909019470215, "load scripts/hypertile_xyz.py": 0.0, "load scripts/model_keyword_support.py": 0.06600236892700195, "load scripts/shared_paths.py": 0.0010004043579101562, "load scripts/tag_autocomplete_helper.py": 0.7999985218048096, "load scripts/!adetailer.py": 1.5400011539459229, "load scripts/clip_interrogator_ext.py": 0.2629988193511963, "load scripts/sd_webui_aspect_ratio_helper.py": 0.38100099563598633, "load scripts/adapter.py": 0.021998167037963867, "load scripts/api.py": 1.600001335144043, "load scripts/batch_hijack.py": 0.0009996891021728516, "load scripts/cldm.py": 0.001001596450805664, "load scripts/controlmodel_ipadapter.py": 0.0009992122650146484, "load scripts/controlnet.py": 0.3750007152557373, "load scripts/controlnet_diffusers.py": 0.0010001659393310547, "load scripts/controlnet_lllite.py": 0.0, "load scripts/controlnet_lora.py": 0.0009989738464355469, "load scripts/controlnet_model_guess.py": 0.0010004043579101562, "load scripts/controlnet_version.py": 0.0, "load scripts/enums.py": 0.0020017623901367188, "load scripts/external_code.py": 0.0, "load scripts/global_state.py": 0.000997781753540039, "load scripts/hook.py": 0.0010006427764892578, "load scripts/infotext.py": 0.0, "load scripts/logging.py": 0.0, "load scripts/lvminthin.py": 0.0010001659393310547, "load scripts/movie2movie.py": 0.0, "load scripts/processor.py": 0.0010006427764892578, "load scripts/utils.py": 0.0, "load scripts/xyz_grid_support.py": 0.0020008087158203125, "load scripts/deepdanbooru_object_recognition.py": 0.10599684715270996, "load scripts/on_app_started.py": 0.13700270652770996, "load scripts/console_log_patch.py": 1.223999261856079, "load scripts/reactor_api.py": 0.2160022258758545, "load scripts/reactor_faceswap.py": 0.006996870040893555, "load scripts/reactor_globals.py": 0.00099945068359375, "load scripts/reactor_helpers.py": 0.0, "load scripts/reactor_logger.py": 0.0010006427764892578, "load scripts/reactor_swapper.py": 0.001001119613647461, "load scripts/reactor_version.py": 0.0, "load scripts/auto.py": 0.0009980201721191406, "load scripts/dino.py": 0.0020020008087158203, "load scripts/process_params.py": 0.00099945068359375, "load scripts/sam.py": 0.11299705505371094, "load scripts/toolkit_gui.py": 0.19000244140625, "load scripts/refiner.py": 0.00099945068359375, "load scripts/seed.py": 0.0, "load scripts": 10.463001251220703, "load upscalers": 0.008000850677490234, "refresh VAE": 0.010998725891113281, "refresh textual inversion templates": 0.0, "scripts list_optimizers": 0.0010008811950683594, "scripts list_unets": 0.0, "reload hypernetworks": 0.02100086212158203, "initialize extra networks": 0.04099774360656738, "scripts before_ui_callback": 0.0019996166229248047, "create ui": 1.823000192642212, "gradio launch": 6.949023485183716, "add APIs": 0.007999897003173828, "app_started_callback/lora_script.py": 0.0, "app_started_callback/tag_autocomplete_helper.py": 0.0029981136322021484, "app_started_callback/clip_interrogator_ext.py": 0.0010001659393310547, "app_started_callback/api.py": 0.009998083114624023, "app_started_callback/on_app_started.py": 4.251001834869385, "app_started_callback/reactor_api.py": 0.004000663757324219, "app_started_callback": 4.268998861312866 } }, "Packages": [ "-rotobuf==3.20.2", "absl-py==2.0.0", "accelerate==0.21.0", "addict==2.4.0", "aenum==3.1.15", "aiofiles==23.2.1", "aiohttp==3.8.5", "aiosignal==1.3.1", "albumentations==1.3.1", "aliyun-python-sdk-alimt==3.2.0", "aliyun-python-sdk-core==2.13.10", "altair==5.1.1", "antlr4-python3-runtime==4.9.3", "anyio==3.7.1", "asttokens==2.4.1", "astunparse==1.6.3", "async-timeout==4.0.3", "attrs==23.1.0", "av==10.0.0", "backcall==0.2.0", "basicsr==1.4.2", "beautifulsoup4==4.12.2", "blendmodes==2022", "boltons==23.0.0", "boto3==1.28.53", "botocore==1.31.53", "cachetools==5.3.1", "certifi==2023.7.22", "cffi==1.15.1", "chardet==5.2.0", "charset-normalizer==3.2.0", "clean-fid==0.1.35", "click==8.1.7", "clip-interrogator==0.6.0", "clip==1.0", "colorama==0.4.6", "coloredlogs==15.0.1", "colorlog==6.8.0", "contourpy==1.1.1", "cryptography==41.0.4", "cssselect2==0.7.0", "cycler==0.11.0", "cython==3.0.2", "decorator==5.1.1", "deprecation==2.1.0", "dill==0.3.7", "easydict==1.10", "einops==0.4.1", "embreex==2.17.7.post4", "exceptiongroup==1.1.3", "executing==2.0.1", "facexlib==0.3.0", "fastapi==0.94.0", "ffmpy==0.3.1", "filelock==3.12.4", "filterpy==1.4.5", "flatbuffers==23.5.26", "fonttools==4.42.1", "frozenlist==1.4.0", "fsspec==2023.9.1", "ftfy==6.1.1", "future==0.18.3", "fvcore==0.1.5.post20221221", "gast==0.5.4", "gdown==4.7.1", "gfpgan==1.3.8", "gitdb==4.0.10", "gitpython==3.1.32", "google-auth-oauthlib==1.0.0", "google-auth==2.23.0", "google-pasta==0.2.0", "gradio-client==0.5.0", "gradio==3.41.2", "grpcio==1.58.0", "h11==0.12.0", "h5py==3.10.0", "httpcore==0.15.0", "httpx==0.24.1", "huggingface-hub==0.17.2", "humanfriendly==10.0", "idna==3.4", "imageio==2.31.3", "importlib-metadata==6.8.0", "importlib-resources==6.1.0", "inflection==0.5.1", "insightface==0.7.3", "iopath==0.1.9", "ipython==8.6.0", "jedi==0.19.1", "jinja2==3.1.2", "jmespath==0.10.0", "joblib==1.3.2", "jsonmerge==1.8.0", "jsonschema-specifications==2023.7.1", "jsonschema==4.19.1", "keras==2.15.0", "kiwisolver==1.4.5", "kornia==0.6.7", "lark==1.1.2", "lazy-loader==0.3", "libclang==16.0.6", "lightning-utilities==0.9.0", "llvmlite==0.41.0", "lmdb==1.4.1", "lpips==0.1.4", "lxml==4.9.3", "mapbox-earcut==1.0.1", "markdown-it-py==3.0.0", "markdown==3.4.4", "markupsafe==2.1.3", "matplotlib-inline==0.1.6", "matplotlib==3.8.0", "mdurl==0.1.2", "mediapipe==0.10.9", "ml-dtypes==0.2.0", "mpmath==1.3.0", "multidict==6.0.4", "multiprocess==0.70.15", "networkx==3.1", "numba==0.58.0", "numpy==1.23.5", "nvidia-cublas-cu11==11.11.3.6", "nvidia-cuda-nvrtc-cu11==11.8.89", "nvidia-cuda-runtime-cu11==11.8.89", "oauthlib==3.2.2", "omegaconf==2.2.3", "onnx-graphsurgeon==0.3.27", "onnx==1.14.1", "onnxruntime-gpu==1.16.1", "onnxruntime==1.15.1", "open-clip-torch==2.20.0", "openai==0.28.0", "opencv-contrib-python==4.8.0.76", "opencv-python-headless==4.8.0.76", "opencv-python==4.8.0.76", "opt-einsum==3.3.0", "orjson==3.9.7", "packaging==23.1", "pandas==2.1.1", "parso==0.8.3", "pathos==0.3.1", "pickleshare==0.7.5", "piexif==1.1.3", "pillow==9.5.0", "pip==23.2.1", "platformdirs==3.10.0", "polygraphy==0.49.0", "portalocker==2.8.2", "pox==0.3.3", "ppft==1.7.6.7", "prettytable==3.9.0", "prompt-toolkit==3.0.41", "protobuf==3.20.3", "psutil==5.9.5", "pure-eval==0.2.2", "py-cpuinfo==9.0.0", "pyasn1-modules==0.3.0", "pyasn1==0.5.0", "pycollada==0.8", "pycparser==2.21", "pydantic==1.10.12", "pydub==0.25.1", "pyexecjs==1.5.1", "pygifsicle==1.0.7", "pygments==2.16.1", "pyparsing==3.1.1", "pyreadline3==3.4.1", "pysocks==1.7.1", "python-dateutil==2.8.2", "python-multipart==0.0.6", "pytorch-lightning==1.9.4", "pytz==2023.3.post1", "pywavelets==1.4.1", "pywin32==306", "pyyaml==6.0.1", "qudida==0.0.4", "realesrgan==0.3.0", "referencing==0.30.2", "regex==2023.8.8", "reportlab==4.0.5", "requests-oauthlib==1.3.1", "requests==2.31.0", "resize-right==0.0.2", "rich==13.5.3", "rpds-py==0.10.3", "rsa==4.9", "rtree==1.1.0", "s3transfer==0.6.2", "safetensors==0.3.1", "scikit-image==0.21.0", "scikit-learn==1.3.1", "scipy==1.11.2", "seaborn==0.12.1", "segment-anything==1.0", "semantic-version==2.10.0", "sentencepiece==0.1.99", "setuptools==68.2.2", "shapely==2.0.2", "six==1.16.0", "smmap==5.0.1", "sniffio==1.3.0", "sounddevice==0.4.6", "soupsieve==2.5", "stack-data==0.6.3", "starlette==0.26.1", "supervision==0.15.0", "support-developer==1.0.5", "svg.path==6.3", "svglib==1.5.1", "sympy==1.12", "tabulate==0.9.0", "tb-nightly==2.15.0a20230922", "tensorboard-data-server==0.7.1", "tensorboard==2.15.1", "tensorflow-estimator==2.15.0", "tensorflow-intel==2.15.0", "tensorflow-io-gcs-filesystem==0.31.0", "tensorflow==2.15.0", "tensorrt-bindings==9.0.1.post11.dev4", "tensorrt-libs==9.0.1.post11.dev4", "tensorrt==9.0.1.post11.dev4", "termcolor==2.3.0", "thop==0.1.1.post2209072238", "threadpoolctl==3.2.0", "tifffile==2023.9.18", "timm==0.9.2", "tinycss2==1.2.1", "tokenizers==0.13.3", "tomesd==0.1.3", "tomli==2.0.1", "toolz==0.12.0", "torch==2.0.1+cu118", "torchdiffeq==0.2.3", "torchmetrics==1.1.2", "torchsde==0.2.6", "torchvision==0.15.2+cu118", "tqdm==4.66.1", "traitlets==5.13.0", "trampoline==0.1.2", "transformers==4.30.2", "trimesh==4.0.8", "typing-extensions==4.8.0", "tzdata==2023.3", "ultralytics==8.0.232", "urllib3==1.26.16", "uvicorn==0.23.2", "wcwidth==0.2.6", "webencodings==0.5.1", "websockets==11.0.3", "werkzeug==2.3.7", "wheel==0.41.2", "wrapt==1.14.1", "xformers==0.0.20", "xxhash==3.4.1", "yacs==0.1.8", "yapf==0.40.1", "yarl==1.9.2", "zipp==3.17.0" ] }

Console logs

** Error completing request                                                                     | 0/20 [00:00<?, ?it/s]
*** Arguments: ('task(tleklbpm7ss5nel)', "A highres photo of a beautiful Irish girl with long voluminous straight brown hair and green eyes, on top of a tall building,with a cityscape behind her.  There's a stunning bokeh effect in the background,", 'fat,ugly,visible hands,daytime,daylight,', [], 15, 'DPM++ 2M Karras', 1, 2, 9, 576, 1032, True, 0.67, 1.25, '4x-UltraSharp', 20, 0, 0, 'realisticVisionV60B1_v50VAE [df24466969]', 'Use same sampler', '', '', ['Face restoration: True'], <gradio.routes.Request object at 0x000001C317181D20>, 0, False, '', 0.8, -1, True, -1, 0.86, 0, 0, False, 'CodeFormer', False, False, {'ad_model': 'face_yolov8n.pt', 'ad_prompt': '', 'ad_negative_prompt': '', 'ad_confidence': 0.3, 'ad_mask_k_largest': 0, 'ad_mask_min_ratio': 0, 'ad_mask_max_ratio': 1, 'ad_x_offset': 0, 'ad_y_offset': 0, 'ad_dilate_erode': 4, 'ad_mask_merge_invert': 'None', 'ad_mask_blur': 4, 'ad_denoising_strength': 0.4, 'ad_inpaint_only_masked': True, 'ad_inpaint_only_masked_padding': 32, 'ad_use_inpaint_width_height': False, 'ad_inpaint_width': 512, 'ad_inpaint_height': 512, 'ad_use_steps': False, 'ad_steps': 28, 'ad_use_cfg_scale': False, 'ad_cfg_scale': 7, 'ad_use_checkpoint': False, 'ad_checkpoint': 'Use same checkpoint', 'ad_use_vae': False, 'ad_vae': 'Use same VAE', 'ad_use_sampler': False, 'ad_sampler': 'DPM++ 2M Karras', 'ad_use_noise_multiplier': False, 'ad_noise_multiplier': 1, 'ad_use_clip_skip': False, 'ad_clip_skip': 1, 'ad_restore_face': False, 'ad_controlnet_model': 'None', 'ad_controlnet_module': 'None', 'ad_controlnet_weight': 1, 'ad_controlnet_guidance_start': 0, 'ad_controlnet_guidance_end': 1, 'is_api': ()}, {'ad_model': 'None', 'ad_prompt': '', 'ad_negative_prompt': '', 'ad_confidence': 0.3, 'ad_mask_k_largest': 0, 'ad_mask_min_ratio': 0, 'ad_mask_max_ratio': 1, 'ad_x_offset': 0, 'ad_y_offset': 0, 'ad_dilate_erode': 4, 'ad_mask_merge_invert': 'None', 'ad_mask_blur': 4, 'ad_denoising_strength': 0.4, 'ad_inpaint_only_masked': True, 'ad_inpaint_only_masked_padding': 32, 'ad_use_inpaint_width_height': False, 'ad_inpaint_width': 512, 'ad_inpaint_height': 512, 'ad_use_steps': False, 'ad_steps': 28, 'ad_use_cfg_scale': False, 'ad_cfg_scale': 7, 'ad_use_checkpoint': False, 'ad_checkpoint': 'Use same checkpoint', 'ad_use_vae': False, 'ad_vae': 'Use same VAE', 'ad_use_sampler': False, 'ad_sampler': 'DPM++ 2M Karras', 'ad_use_noise_multiplier': False, 'ad_noise_multiplier': 1, 'ad_use_clip_skip': False, 'ad_clip_skip': 1, 'ad_restore_face': False, 'ad_controlnet_model': 'None', 'ad_controlnet_module': 'None', 'ad_controlnet_weight': 1, 'ad_controlnet_guidance_start': 0, 'ad_controlnet_guidance_end': 1, 'is_api': ()}, UiControlNetUnit(enabled=False, module='none', model='None', weight=1, image=None, resize_mode='Crop and Resize', low_vram=False, processor_res=-1, threshold_a=-1, threshold_b=-1, guidance_start=0, guidance_end=1, pixel_perfect=False, control_mode='Balanced', inpaint_crop_input_image=False, hr_option='Both', save_detected_map=True, advanced_weighting=None), UiControlNetUnit(enabled=False, module='none', model='None', weight=1, image=None, resize_mode='Crop and Resize', low_vram=False, processor_res=-1, threshold_a=-1, threshold_b=-1, guidance_start=0, guidance_end=1, pixel_perfect=False, control_mode='Balanced', inpaint_crop_input_image=False, hr_option='Both', save_detected_map=True, advanced_weighting=None), UiControlNetUnit(enabled=False, module='none', model='None', weight=1, image=None, resize_mode='Crop and Resize', low_vram=False, processor_res=-1, threshold_a=-1, threshold_b=-1, guidance_start=0, guidance_end=1, pixel_perfect=False, control_mode='Balanced', inpaint_crop_input_image=False, hr_option='Both', save_detected_map=True, advanced_weighting=None), None, True, '0', '0', 'inswapper_128.onnx', 'CodeFormer', 1, True, '8x_NMKD-Faces_160000_G', 4.4, 1, False, True, 1, 0, 0, False, 0.5, True, False, 'CPU', True, 1, 'Kelly2.safetensors', '', None, False, False, 0, None, [], 0, False, [], [], False, 0, 1, False, False, 0, None, [], -2, False, [], False, 0, None, None, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, 0, False, None, None, False, None, None, False, None, None, False, 50) {}
    Traceback (most recent call last):
      File "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py", line 57, in f
        res = list(func(*args, **kwargs))
      File "C:\AI\Image AI\SDNEW\webui\modules\call_queue.py", line 36, in f
        res = func(*args, **kwargs)
      File "C:\AI\Image AI\SDNEW\webui\modules\txt2img.py", line 55, in txt2img
        processed = processing.process_images(p)
      File "C:\AI\Image AI\SDNEW\webui\modules\processing.py", line 734, in process_images
        res = process_images_inner(p)
      File "C:\AI\Image AI\SDNEW\webui\extensions\sd-webui-controlnet\scripts\batch_hijack.py", line 42, in processing_process_images_hijack
        return getattr(processing, '__controlnet_original_process_images_inner')(p, *args, **kwargs)
      File "C:\AI\Image AI\SDNEW\webui\modules\processing.py", line 857, in process_images_inner
        p.setup_conds()
      File "C:\AI\Image AI\SDNEW\webui\modules\processing.py", line 1308, in setup_conds
        super().setup_conds()
      File "C:\AI\Image AI\SDNEW\webui\modules\processing.py", line 470, in setup_conds
        self.c = self.get_conds_with_caching(prompt_parser.get_multicond_learned_conditioning, prompts, total_steps, [self.cached_c], self.extra_network_data)
      File "C:\AI\Image AI\SDNEW\webui\modules\processing.py", line 455, in get_conds_with_caching
        cache[1] = function(shared.sd_model, required_prompts, steps, hires_steps, shared.opts.use_old_scheduling)
      File "C:\AI\Image AI\SDNEW\webui\modules\prompt_parser.py", line 261, in get_multicond_learned_conditioning
        learned_conditioning = get_learned_conditioning(model, prompt_flat_list, steps, hires_steps, use_old_scheduling)
      File "C:\AI\Image AI\SDNEW\webui\modules\prompt_parser.py", line 188, in get_learned_conditioning
        conds = model.get_learned_conditioning(texts)
      File "C:\AI\Image AI\SDNEW\webui\modules\sd_models_xl.py", line 31, in get_learned_conditioning
        c = self.conditioner(sdxl_conds, force_zero_embeddings=['txt'] if force_zero_negative_prompt else [])
      File "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py", line 1538, in _call_impl
        result = forward_call(*args, **kwargs)
      File "C:\AI\Image AI\SDNEW\webui\repositories\generative-models\sgm\modules\encoders\modules.py", line 141, in forward
        emb_out = embedder(batch[embedder.input_key])
      File "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
        return forward_call(*args, **kwargs)
      File "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_clip.py", line 234, in forward
        z = self.process_tokens(tokens, multipliers)
      File "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_clip.py", line 273, in process_tokens
        z = self.encode_with_transformers(tokens)
      File "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack_open_clip.py", line 57, in encode_with_transformers
        d = self.wrapped.encode_with_transformer(tokens)
      File "C:\AI\Image AI\SDNEW\webui\repositories\generative-models\sgm\modules\encoders\modules.py", line 467, in encode_with_transformer
        x = self.model.token_embedding(text)  # [batch_size, n_ctx, d_model]
      File "C:\AI\Image AI\SDNEW\system\python\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
        return forward_call(*args, **kwargs)
      File "C:\AI\Image AI\SDNEW\webui\modules\sd_hijack.py", line 359, in forward
        tensor = torch.cat([tensor[0:offset + 1], emb[0:emb_len], tensor[offset + 1 + emb_len:]])
    RuntimeError: Sizes of tensors must match except in dimension 0. Expected size 1280 but got size 768 for tensor number 1 in the list.

Additional information

No response

ipinz commented 8 months ago

Wrong version when using of lora? SDXL or SD1.5? In the log it shows something wrong with the embedding

Screenshot 2024-01-11 075059

elfgoose commented 8 months ago

Thanks for the reply. It's weird because I'm not using a lora or an embedding on this generation. I'm using ReActor, but I'm also using it on the ones that work fine

harpy2 commented 6 months ago

same here

bigmover commented 1 month ago

Thanks for the reply. It's weird because I'm not using a lora or an embedding on this generation. I'm using ReActor, but I'm also using it on the ones that work fine

Have you resolve the problem? Would you mind to share how to resolve it ?