I tried training the coreference model (using the instructions in the README), but the training job was killed due to infra issues after the training for 17 full epochs (out of 19 requested epochs).
I'd like to train this model to completion, to reproduce the results in the paper, but would like to avoid spending the GPU resources to re-train the first 17 epochs again. Is there a way to do this with the allennlp training scheme?
I tried training the coreference model (using the instructions in the README), but the training job was killed due to infra issues after the training for 17 full epochs (out of 19 requested epochs).
I'd like to train this model to completion, to reproduce the results in the paper, but would like to avoid spending the GPU resources to re-train the first 17 epochs again. Is there a way to do this with the
allennlp
training scheme?