Open qibao77 opened 4 years ago
Adding the KL loss for latent codes is only used to validate the effects of loss functions in my exps, e.g., jointing gan loss and the KL loss, but the results in my exps show it has a little effect for metrics. You could also remove it from the source codes.
Thank you for your reply!
Adding the KL loss for latent codes is only used to validate the effects of loss functions in my exps, e.g., jointing gan loss and the KL loss, but the results in my exps show it has a little effect for metrics. You could also remove it from the source codes.
Another problem, I found that the KL loss in your code is actually L2 regularization (only using the mean), but the KL loss of VAE should include mean and var. Why is there such a difference?
I have the same question about the KL loss,. It seems like the author only use the L2 to make the mean to zero, which is different from the regular KL divergence. So can you explain the difference?
The work is interesting! However, in your paper, you only add a KL divergence loss to regularize the distribution of the noise code, while you add KL loss to all latent features in your open source code. Why is there such a difference? Is KL loss important to the final result?