Closed krisrs1128 closed 4 years ago
@krisrs1128 did you see I added a --resume
flag some time ago and a parallel_resume.py
file?
Yep, I'm basically just using the same code, but letting you specify the model, rather than looking up what's in the written conf. I'm guessing we may want to initialize using a model from a different run.
The changes related to extrapolation / step in extragradient are because, upon resume, I think it might try to take a step before extrapolation (depending on the number of steps in the checkpoint).
makes sense sorry :)
This creates an
init_chckpt_dir
andinit_chkpt_step
which let's you initialize with a previous checkpoint, using the existingresume()
method.In theory, you could fake the checkpointing and place arbitrary pretrained models in the directory.