Open daidaiershidi opened 6 days ago
Thank you again for your interest in our work.
I think the script you are using does not have big issues.
However, since you are a beginner and this is your first time training an LLM, I recommend you directly try training utilizing the original Alpaca
dataset, which is provided in our data
folder, and see if the performance is reasonable. By doing this you can get to know if the training pipeline you built is correct or not.
| Avg | ARC-challenge | HellaSwag_acc_norm | MMLU | truthfulqa_mc2 -- | -- | -- | -- | -- | -- Alpaca in paper | 50.21 | 42.65 | 76.91 | 41.73 | 39.55 WizardLM-10% in paper | 57.57 | 54.86 | 80.46 | 45.74 | 49.20 my_alpaca | 52.62 | 46.25 | 76.18 | 45.39 | 42.65 my_WizardLM-10% | 52.08 | 46.67 | 74.70 | 40.00 | 46.94Thank you again for your interest in our work.
I think the script you are using does not have big issues. However, since you are a beginner and this is your first time training an LLM, I recommend you directly try training utilizing the original
Alpaca
dataset, which is provided in ourdata
folder, and see if the performance is reasonable. By doing this you can get to know if the training pipeline you built is correct or not.
It seems that training with the full alpaca dataset is no problem, but using the 10% WizardLM training llama2-7b from Data and Model Weights V1, the performance is worse
Thank you very much for the work you have brought, which is very helpful for those of us with fewer training resources. I am a newcomer to the field of NLP and am not very familiar with training frameworks (this is my first time training an LLM), and I find that my final results are very poor, even far worse than another reproducer. Here is the script I used:
When training the final model:
I would like to ask if there is a mistake in my training configuration?