Open jbmaxwell opened 3 months ago
someone actually pull requested this in and i'm unfamiliar with it
does the reconstruction loss look good?
Well, actually, I've just discovered that it seems to be an LR thing...
zooming in on first 10k steps:
But recon seems to converge pretty steadily. So maybe just a false alarm. I still have to wrap my head around how LatentQuantize works (and how to get what I want), mind you!
I'm trying to use the LatentQuantize model in an autoencoder context. My inputs are flat 1-d tensors (32) and my encoder passes a shape of (batch_size, 64) to the quantizer. For now, my "levels" is [8, 6, 4], my latent_dim is 64:
The loss starts at zero, then exponentially increases:
Any thoughts as to why this might happen?