openvax / mhcflurry

Peptide-MHC I binding affinity prediction
http://openvax.github.io/mhcflurry/
Apache License 2.0
193 stars 58 forks source link

Learning rate matters #64

Closed iskandr closed 6 years ago

iskandr commented 8 years ago

We're currently using adaptive SGD (mostly RMSprop) and hoping that the default parameters work for us. Many recent (and not so recent) deep learning papers have a schedule for decreasing the learning rate over time, often phrased something like "halve the learning rate every 50 epochs". Here's an extreme case from Densely Connected Convolutional Networks (learning rate drop by 10x accompanied with major decrease in loss):

DenseNet loss vs. epochs

I imagine the following three parameters would be a useful addition to the cross-validation loop used to determine optimal hyperparameters:

timodonnell commented 6 years ago

I played around with learning rate informally and did not see opportunities for easy wins, although I didn't explore the full space of decay schedules etc. Closing for now.