Open JasonLLLLLLLLLLL opened 1 year ago
Where does your checkpoint come from? If you start a fine-tune, checkpoints are created automatically and if you then interrupt the fine-tuning and start it again (with the same --output
directory), training should be resumed from the latest checkpoint automatically.
It seems it can get the last checkpoint in train/llama.py. But the loss seems to start over again(at 1.6). It should be 0.2 at this checkpoint.
or can I code like this in train/llama.py?
sorry to bother you for those questions. I am new to LLM fitune. I hope I can get your answer.