Closed menglin0320 closed 7 years ago
No philosophy behind any of this, it's just a kind of starting point if anybody wants to try to experiment around. I believe that I found that the large epsilon learned a little faster, and I think that the reason I tried it was because of ikostrikov's implementation (which has since changed significantly): https://github.com/ikostrikov/TensorFlow-VAE-GAN-DRAW
And you use elu as the activation function, is this activation function guaranteed to be better than relu for this task?