gauss5930 / AlpaGasus2-QLoRA

This is AlpaGasus2-QLoRA based on LLaMA2 with AlpaGasus mechanism using QLoRA!
14 stars 2 forks source link

What is the difference between Alpagasus-2-13b-QLoRA-merged and Alpagasus-2-13b-QLoRA-pipeline ? #2

Open MDK-L opened 6 months ago

MDK-L commented 6 months ago

image I found these two models on OpenAI's LLM, and they exhibit significant differences in performance. I'm wondering why? Could it be due to differences in floating-point precision?

gauss5930 commented 6 months ago

Hello @MDK-L! Thank you for your interest in our AlpaGasus2-QLoRA.

As you asked, you can see that two AlpaGasus2-QLoRA models have been uploaded to the Open LLM Leaderboard. Among the two models, AlpaGasus-2-13b-QLoRA-pipeline is the model created for testing in the initial experiments. The final model created after additional modifications was AlpaGasus-2-13b-QLoRA-merged. The difference in performance between the two models is probably due to differences in the learning process.

I hope this answer helps resolve your questions, and if you have any additional questions, please feel free to ask.

MDK-L commented 6 months ago
1702624462678

Thank you for your answer, but I still have some doubts. When I choose the floating point precision in Open LLM Leaderboard, the 4-bit corresponds to a "pipeline" model with lower accuracy, while float16 and bfloat16 correspond to two "merged" models with similar higher accuracy. This confuses me. Shouldn't the Qlora model be 4-bit?