Closed Elekez closed 1 year ago
after install animatediff and download model, I try to run it but got errors
should generate gif
webui: A1111 1.6 extension: newest
No response
-xformers --upcast-sampling --no-hashing --always-batch-cond-uncond --medvram --precision full --no-half
*** Error completing request *** Arguments: ('task(m56nl9t188jur0f)', '', '', [], 20, 'Euler a', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same che ckpoint', 'Use same sampler', '', '', [], <gradio.routes.Request object at 0x000001D56FEF4970>, 0, False, '', 0.8, -1, False, -1, 0, 0 , 0, <scripts.animatediff_ui.AnimateDiffProcess object at 0x000001D571065210>, False, False, 'positive', 'comma', 0, False, False, '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, 0, False) {} Traceback (most recent call last): File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\call_queue.py", line 57, in f res = list(func(*args, **kwargs)) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\call_queue.py", line 36, in f res = func(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\txt2img.py", line 55, in txt2img processed = processing.process_images(p) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\processing.py", line 732, in process_images res = process_images_inner(p) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\processing.py", line 867, in process_images_inner samples_ddim = p.sample(conditioning=p.c, unconditional_conditioning=p.uc, seeds=p.seeds, subseeds=p.subseeds, subseed_strengt h=p.subseed_strength, prompts=p.prompts) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\processing.py", line 1140, in sample samples = self.sampler.sample(self, x, conditioning, unconditional_conditioning, image_conditioning=self.txt2img_image_conditi oning(x)) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\sd_samplers_kdiffusion.py", line 235, in sample samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args=self.sampler_extra_args, disable=Fa lse, callback=self.callback_state, **extra_params_kwargs)) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\sd_samplers_common.py", line 261, in launch_sampling return func() File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\sd_samplers_kdiffusion.py", line 235, in <lambda> samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args=self.sampler_extra_args, disable=Fa lse, callback=self.callback_state, **extra_params_kwargs)) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorat e_context return func(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\k-diffusion\k_diffusion\sampling.py", line 145, in sample_e uler_ancestral denoised = model(x, sigmas[i] * s_in, **extra_args) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\scripts\animatediff_infv2v.py", line 274 , in mm_cfg_forward x_out = mm_sd_forward(self, x_in, sigma_in, cond_in, image_cond_in, make_condition_dict) # hook File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\scripts\animatediff_infv2v.py", line 188 , in mm_sd_forward out = self.inner_model(x_in[_context], sigma_in[_context], cond=make_condition_dict(cond_in[_context], image_cond_in[_context] )) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\k-diffusion\k_diffusion\external.py", line 112, in forward eps = self.get_eps(input * c_in, self.sigma_to_t(sigma), **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\k-diffusion\k_diffusion\external.py", line 138, in get_eps return self.inner_model.apply_model(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\sd_hijack_utils.py", line 17, in <lambda> setattr(resolved_obj, func_path[-1], lambda *args, **kwargs: self(*args, **kwargs)) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\sd_hijack_utils.py", line 28, in __call__ return self.__orig_func(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py" , line 858, in apply_model x_recon = self.model(x_noisy, t, **cond) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1538, in _call_ impl result = forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py" , line 1335, in forward out = self.diffusion_model(x, t, context=cc) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1538, in _call_ impl result = forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\modules\sd_unet.py", line 91, in UNetModel_forward return ldm.modules.diffusionmodules.openaimodel.copy_of_UNetModel_forward_for_webui(self, x, timesteps, context, *args, **kwar gs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\ openaimodel.py", line 797, in forward h = module(h, emb, context) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\ openaimodel.py", line 86, in forward x = layer(x) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\motion_module.py", line 86, in forward return self.temporal_transformer(input_tensor, encoder_hidden_states, attention_mask) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\motion_module.py", line 150, in forward hidden_states = block(hidden_states, encoder_hidden_states=encoder_hidden_states, video_length=video_length) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\motion_module.py", line 212, in forward hidden_states = attention_block( File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_ impl return forward_call(*args, **kwargs) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\motion_module.py", line 567, in forward hidden_states = self._memory_efficient_attention(query, key, value, attention_mask, optimizer_name) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\extensions\sd-webui-animatediff\motion_module.py", line 467, in _memory_ efficient_attention hidden_states = xformers.ops.memory_efficient_attention( File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\xformers\ops\fmha\__init__.py", line 192, in memo ry_efficient_attention return _memory_efficient_attention( File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\xformers\ops\fmha\__init__.py", line 290, in _mem ory_efficient_attention return _memory_efficient_attention_forward( File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\xformers\ops\fmha\__init__.py", line 310, in _mem ory_efficient_attention_forward out, *_ = op.apply(inp, needs_gradient=False) File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\xformers\ops\fmha\cutlass.py", line 175, in apply out, lse, rng_seed, rng_offset = cls.OPERATOR( File "E:\pinokio\api\sd-webui.pinokio.git\automatic1111\venv\lib\site-packages\torch\_ops.py", line 502, in __call__ return self._op(*args, **kwargs or {}) RuntimeError: CUDA error: invalid configuration argument CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
Is there an existing issue for this?
Have you read FAQ on README?
What happened?
after install animatediff and download model, I try to run it but got errors
Steps to reproduce the problem
What should have happened?
should generate gif
Commit where the problem happens
webui: A1111 1.6 extension: newest
What browsers do you use to access the UI ?
No response
Command Line Arguments
Console logs
Additional information
No response