Our initial experiments have shown that our model trains very slowly.
Thus, we made our hyperparameter search train for a large number of epochs.
However, some hyperparameter selections are very bad and we can know after a few epochs (e.g. 5 or so) that the model has converged at some really bad score.
Thus, we shouldn't bother training for 40 epochs if we know after 8 that the model isn't improving.
This task is to make it so that our hyperparameter search gives up early in these cases.
Our initial experiments have shown that our model trains very slowly.
Thus, we made our hyperparameter search train for a large number of epochs.
However, some hyperparameter selections are very bad and we can know after a few epochs (e.g. 5 or so) that the model has converged at some really bad score.
Thus, we shouldn't bother training for 40 epochs if we know after 8 that the model isn't improving.
This task is to make it so that our hyperparameter search gives up early in these cases.