yookoon / VLAE

Code for Variational Laplace Autoencoders
MIT License
54 stars 7 forks source link

why is the generated images so bad? #2

Closed zh-ding closed 5 years ago

zh-ding commented 5 years ago

image

yookoon commented 5 years ago

Will you try smaller latent variable dimensions? (e.x. 16)

We have experienced similar issues when the latent variable dimension gets large on Gaussian output models. We hypothesize that it happens when the latent variable dimension exceeds the intrinsic data manifold dimension, so that some latent dimensions fit to noise dimensions of data.

Or you can use Bernoulli output models, which are free from such artifacts.

zh-ding commented 5 years ago

yeah, been better, but still much worse than the vanilla vae

yookoon commented 5 years ago

It will be helpful if you provide more details. Else, I will close this issue.

zh-ding commented 5 years ago

yeah thanks for your reply. I've changed the z-dim smaller to 20 as you said before, but the generated results are still not good as vanilla vae. Maybe I've misunderstood your paper, but from my perspective, you are trying to use local linearity to fit global non-linearity?

yookoon commented 5 years ago

If you worry about the visual generation quality on MNIST, I suggest you use Bernoulli output models.

I'm not sure what you mean by global non-linearity, but the gist of our work is that we use local linearity of ReLU networks to get a better posterior approximation for latent variable models. Of course, the data manifold will be non-linear in general, so we are modeling the nonlinear manifold with a piece-wise linear one.

zh-ding commented 5 years ago

Oh I see, very appreciate for your help.