Closed B-201 closed 2 weeks ago
I think some fluctuations is reasonable. We do have some update checkpoints coming soon trained with large batch size.
I’m curious why the index of full train is not as good as lora?
We don't know either. One hypothesis is that the full finetune will overfit to training set more easily.
I tested some of the datasets in the eval set and found that the results are higher than those in the paper. Is there an update to the LoRA model compared to what was presented in the paper?