Closed sampepose closed 8 years ago
Hi. You should definitely have another nonlinearity in the last layer. My suggestion is to use sigmoid and to normalize the target to [0, 1]. If that doesn't help, I would play with different learning rates.
I'd be curious to see what the architecture looks like. And if Op ever figured out what was wrong with his example. @BenjaminBossan's advices likely to help
Theano has documentation on Dealing with NaNs which should also help.
Hi! I'm trying to make a convolutional autoencoder based off of VGG-S (https://github.com/Lasagne/Recipes/blob/master/modelzoo/vgg_cnn_s.py).
For some reason, learning always converges to NaN almost immediately. I think my architecture is correct from VGG-S, so I'm not sure why this is happening.
Here's my code (https://gist.github.com/sampepose/ccb58557271cff10d182f4ab8282f3b4).