bst-mug / n2c2

Support code for participation at the 2018 n2c2 Shared-Task Track 1
https://n2c2.dbmi.hms.harvard.edu
Apache License 2.0
6 stars 4 forks source link

Plot test stats over number of epochs #101

Closed michelole closed 5 years ago

michelole commented 5 years ago

Hypothesis: the more epochs, the more we overfit.

michelole commented 5 years ago

tl;dr: fasttext with pre-trained embeddings indeed converges faster (less epochs) to the max accuracy. This effect is stronger the larger the learning rate. Anyway, fasttext with self-trained embeddings always seem to catch up at some point.