Closed alecGraves closed 5 years ago
Hi,
I'm using your implementation of the beta-VAE. You haven't committed this change of the loss function to your code. Should we use the new one you proposed or the old one? Have you tested that the new one works and does what it is supposed to do?
Thank you in advance!
I have not tested anything. I would recommend leaving the capacity argument at the default value of zero for now.
Ok, thank you for your reply! In the beta-VAE paper (https://openreview.net/references/pdf?id=Sy2fzU9gl) there's no capacity parameter. They just play with beta and the size of the latent representation, am I right?
Another suggestion would be to normalize the beta, as they do in the paper, to take into account the dimension of the input and dimension of the latent space :-)
@beatriz-ferreira Thank you for the suggestion, I missed that detail from the paper, but normalization makes a lot of sense.
I removed the capacity argument from the layer. The idea behind the capacity argument is to guide the model to learn a representation with a specific number of distributions, but still allow the model to expand the number used if necessary. It is kinda silly, as a standard VAE allows one to specify the number of distributions to use directly.
I removed this (nonworking) argument from the sampling layer in a recent commit.
Closing...
What was past me thinking? I do not know 😄
I just randomly subtract a constant from my loss?
This is more like it: