uclnlp / jack

Jack the Reader
MIT License
257 stars 82 forks source link

Learning rate decay too high by default #322

Closed TimDettmers closed 6 years ago

TimDettmers commented 6 years ago

It is unclear for users that learning rate decay is active. I would argue it should be set to 0 by default, 0.5 can be pretty high. Training can be noisy in the beginning if one starts from random weights and the default learning rate decay behavior can destroy a training procedure easily.

dirkweissenborn commented 6 years ago

alright, but it should be 1 instead, because it is exponential decay.