Open xiankgx opened 2 years ago
+1 same problem here.
Try to decrease your batch size, too big batch size may cause this problem and if it is too big it will be nan. Reduce the batch size to 1 4 8 works for me.
+1 same problem here, do you solve it
reducing the learning rate is working for me
Hi, I took a first stage KL-regularized autoencoder from one of the pretained models (models/first_stage_models/kl-f8/model.ckpt) and tried to train a LDM model. Training proceeds, but the loss ((train/loss_simple_step)) hardly budge. From the image, I'm seeing good quality reconstructions (due to the pretrained autoencoder). However, the samples are just noise, something you would expect if you randomize the latent code in VQGAN.