kabachuha / sd-webui-text2video

Auto1111 extension implementing text2video diffusion models (like ModelScope or VideoCrafter) using only Auto1111 webui dependencies
Other
1.28k stars 107 forks source link

[Bug]: #78

Closed GrishDaProgrammer closed 1 year ago

GrishDaProgrammer commented 1 year ago

Is there an existing issue for this?

Are you using the latest version of the extension?

What happened?

All txt2img generations resulting in:

Exception occurred: argument of type 'NoneType' is not iterable

I have triple checked the installation instructions and re-downloaded the models, in an attempt to prevent simple oversight.

Steps to reproduce the problem

  1. Click Generate using following settings (or any other combination of settings that I have tried)

image

  1. Pipeline launches successfully.
  2. DDIM sampling tensor(1) completes sucessfully.
  3. Following Fault Occurs

ModelScope text2video extension for auto1111 webui Git commit: c8335de9 (Mon Mar 27 19:49:45 2023) Starting text2video Pipeline setup config namespace(framework='pytorch', task='text-to-video-synthesis', model={'type': 'latent-text-to-video-synthesis', 'model_args': {'ckpt_clip': 'open_clip_pytorch_model.bin', 'ckpt_unet': 'text2video_pytorch_model.pth', 'ckpt_autoencoder': 'VQGAN_autoencoder.pth', 'max_frames': 16, 'tiny_gpu': 1}, 'model_cfg': {'unet_in_dim': 4, 'unet_dim': 320, 'unet_y_dim': 768, 'unet_context_dim': 1024, 'unet_out_dim': 4, 'unet_dim_mult': [1, 2, 4, 4], 'unet_num_heads': 8, 'unet_head_dim': 64, 'unet_res_blocks': 2, 'unet_attn_scales': [1, 0.5, 0.25], 'unet_dropout': 0.1, 'temporal_attention': 'True', 'num_timesteps': 1000, 'mean_type': 'eps', 'var_type': 'fixed_small', 'losstype': 'mse'}}, pipeline={'type': 'latent-text-to-video-synthesis'}) device cuda Working in txt2vid mode 0%| | 0/1 [00:00<?, ?it/s]latents torch.Size([1, 4, 24, 32, 32]) tensor(0.0008, device='cuda:0') tensor(1.0006, device='cuda:0') DDIM sampling tensor(1): 100%|██████████████████████████████████████████████████████████████████████████████████| 31/31 [00:43<00:00, 1.40s/it] Traceback (most recent call last):██████████████████████████████████████████████████████████████████████████████| 31/31 [00:43<00:00, 1.30s/it] File "C:\Users\XXX\stable-diffusion-webui\extensions\sd-webui-modelscope-text2video\scripts\modelscope-text2vid.py", line 160, in process samples, = pipe.infer(prompt, n_prompt, steps, frames, seed + batch if seed != -1 else -1, cfg_scale, File "C:\Users\XXX\stable-diffusion-webui\extensions\sd-webui-modelscope-text2video\scripts\t2v_pipeline.py", line 255, in infer if 'CPU' in cpu_vae: TypeError: argument of type 'NoneType' is not iterable Exception occurred: argument of type 'NoneType' is not iterable

What should have happened?

No response

WebUI and Deforum extension Commit IDs

webui commit id - 955df77 txt2vid commit id -c8335de9

What GPU were you using for launching?

3060 -12 gb

On which platform are you launching the webui backend with the extension?

No response

Settings

image

Console logs

ModelScope text2video extension for auto1111 webui
Git commit: c8335de9 (Mon Mar 27 19:49:45 2023)
Starting text2video
Pipeline setup
config namespace(framework='pytorch', task='text-to-video-synthesis', model={'type': 'latent-text-to-video-synthesis', 'model_args': {'ckpt_clip': 'open_clip_pytorch_model.bin', 'ckpt_unet': 'text2video_pytorch_model.pth', 'ckpt_autoencoder': 'VQGAN_autoencoder.pth', 'max_frames': 16, 'tiny_gpu': 1}, 'model_cfg': {'unet_in_dim': 4, 'unet_dim': 320, 'unet_y_dim': 768, 'unet_context_dim': 1024, 'unet_out_dim': 4, 'unet_dim_mult': [1, 2, 4, 4], 'unet_num_heads': 8, 'unet_head_dim': 64, 'unet_res_blocks': 2, 'unet_attn_scales': [1, 0.5, 0.25], 'unet_dropout': 0.1, 'temporal_attention': 'True', 'num_timesteps': 1000, 'mean_type': 'eps', 'var_type': 'fixed_small', 'loss_type': 'mse'}}, pipeline={'type': 'latent-text-to-video-synthesis'})
device cuda
Working in txt2vid mode
  0%|                                                                                                                     | 0/1 [00:00<?, ?it/s]latents torch.Size([1, 4, 24, 32, 32]) tensor(0.0008, device='cuda:0') tensor(1.0006, device='cuda:0')
DDIM sampling tensor(1): 100%|██████████████████████████████████████████████████████████████████████████████████| 31/31 [00:43<00:00,  1.40s/it]
Traceback (most recent call last):██████████████████████████████████████████████████████████████████████████████| 31/31 [00:43<00:00,  1.30s/it]
  File "C:\Users\xxx\stable-diffusion-webui\extensions\sd-webui-modelscope-text2video\scripts\modelscope-text2vid.py", line 160, in process
    samples, _ = pipe.infer(prompt, n_prompt, steps, frames, seed + batch if seed != -1 else -1, cfg_scale,
  File "C:\Users\xxx\stable-diffusion-webui\extensions\sd-webui-modelscope-text2video\scripts\t2v_pipeline.py", line 255, in infer
    if 'CPU' in cpu_vae:
TypeError: argument of type 'NoneType' is not iterable
Exception occurred: argument of type 'NoneType' is not iterable

Additional information

Windows 10 running locally.

hithereai commented 1 year ago

Tried, but couldn't replicate unfortunately.

Windows 11, same git commits as yours.

hithereai commented 1 year ago

How this looks at yours?

image

From webui settings -->ModelScopeTxt2Vid section.

GrishDaProgrammer commented 1 year ago

Same settings, GPU on half precision and not saving the model to VRAM.

Compviztr commented 1 year ago

Also getting the same error. Was working fine before latest update. Windows 11

hithereai commented 1 year ago

try to delete those 2 files from your root of webui folder: image

then relaunch webui and retest.

Compviztr commented 1 year ago

Same error after deleting above files and restarting

Compviztr commented 1 year ago

It works now after going to the settings and checking “keep model in VRAM…..”

GrishDaProgrammer commented 1 year ago

It works now after going to the settings and checking “keep model in VRAM…..”

Thanks!

The combo of removing (then restarting A111) config.json ui-config.json

&

Running with Keep model in VRAM checked avoids the bug.