-
sorry, i cant find the model in the https://cloud.tsinghua.edu.cn/f/7eb11fc208144ed0ad20/?dl=1, and it shows "对不起,您访问的页面不存在", could you please share the pretrained model? thank you very much!
-
When i ran model on imagenet: one error happened:
How can i tackle the problem about tensor? Dashen / masters help me , thanks!
> Model unexpected keys:
['transformer.log_alpha', 'transformer.log…
-
I realize that the configuration of VQ autoencoders in Latent Diffusion is different than the one used in VQGAN (taming-transformers). Specifically, I see that **embed_dim** and **z_channels** have lo…
-
Hello, I wanted to ask if there is a possibility to have VQ-BeT running on multiple camera's for some environments that have different views, like Robomimic? If so can someone give me points on what e…
-
Hi,
When I am trying to train a LDM model as "CUDA_VISIBLE_DEVICES=1 python main.py --base configs/latent-diffusion/lsun_bedrooms-ldm-vq-4.yaml -t --gpus 0", there is an error massage as "No such f…
-
Why does the Latent Diffusion Model use **variational autoencoders (VAE)** or similar generative models like **VQ-GAN/VAE** for compression instead of using **AutoEncoder (AE)?** If AE can be consider…
-
$ python inference.py -i testdata/RealSet65 -o results/SinSR/RealSet65 --scale 4 --ckpt weights/SinSR_v1.pth --one_step
UserWarning: The torchvision.transforms.functional_tensor module is deprecated …
-
DiVAE [1] uses a VQ encoder and a diffusion decoder. Unfortunately, there's no public implementation. It would also be nice to combine that with diffusion Transformers [2].
Any way many thanks for …
-
Hi,
Thanks for providing this really convenient package to use the CLIP model!
I've come across a problem with `build_model` when trying to reconstruct the model from a state_dict on my local c…
-
hey @adelacvg thank for sharing the code
after reading the code i want to ask you few question about new 24k model if you dont mind
1. what make different about this model from previous one (ht…
acul3 updated
1 month ago