gangchill / nip-convnet

Convolutional Autoencoder implemented in Tensorflow
Apache License 2.0
2 stars 3 forks source link

Deep Autoencoder doesn't work with ReLU activation function #8

Open maxkohlbrenner opened 7 years ago

maxkohlbrenner commented 7 years ago

A deeper autoencoder only learns something useful with a sigmoid activation for instance, for the weight transfer to the CNN, a relu activation would e more useful. Find out why it doesn't work.

maxkohlbrenner commented 7 years ago

For a simple setting, all relu units die very quickly relu_problem_of_death


AUTOENCODER SPECIFICATIONS filter_dims = [(5,5)] hidden_channels = [5] use_max_pooling = False strides = None # other strides should not work yet activation_function = 'relu' batch_size = 100 max_iterations = 50 chk_iterations = 10 step_size = 0.0001