-
Hi, great work!
But i am confused that why do u encode motion using motion VQVAE ? As far as i know ,many works aims to generate motion and it is convenient to use the decoder of VQVAE . But MotionLL…
-
hi, great work!
I was wondering if there is an evaluation code for the VQVAE.
I tested with another codebase, and I could not get the same results as the paper.
It will be great if you can s…
-
I am trying to quantize an image into a tensor of indices, then decode from it, but I am getting float latents.
My full code:
```py
from huggingface_hub import hf_hub_download
from diffusers import V…
-
Thanks for sharing your excellent work.
I'm trying to replicate the results but encountered some problems. According to your code, the MaskedTransformer predicts the base tokens encoded by VQVAE, an…
-
- This is the reconstruction result of 100 Epoch of VQVAE.
![recons_VQVAE_Epoch_99](https://user-images.githubusercontent.com/27271981/204431756-8dd9346a-5ed3-4560-951c-558f1f31e455.png)
-
Thank you so much for sharing!
Could you provide insights into the number of epochs required to achieve high-resolution, fine details during VQVAE training for 256x256 RGB images?
Additionally, …
-
Hello. Did all as described in manual.
Python 3.8
pip install -r requirements.txt all fine
Loadaded vgg.pth
Prepared mnist dataset.
But while training VQVAE got this error
(.conda) (bas…
-
-
Dear author, in your fine-tuning code, I found that vqvae is not used. Why?
-
Hello, can you tell me more about the training details of the VQVAE, for example, the training dataset, batchsize and number of epoches, these are not clearly written in the Paper, beside, for the VQ…