Open fzyzcjy opened 3 weeks ago
Thank you for your question, I will check and reply soon.
fzyzcjy @.***>于2024年10月26日 周六13:02写道:
Hi thanks for the library! https://github.com/unslothai/unsloth is a library that supports fast PEFT fine-tuning. Therefore, I wonder whether this is / will be compatible with that?
— Reply to this email directly, view it on GitHub https://github.com/xichen-fy/Fira/issues/2, or unsubscribe https://github.com/notifications/unsubscribe-auth/BLY3UD4HO6KJFH5BJQINONDZ5MO6XAVCNFSM6AAAAABQUOIF7SVHI2DSMVQWIX3LMV43ASLTON2WKOZSGYYTKNJVGAZTANA . You are receiving this because you are subscribed to this thread.Message ID: @.***>
In principle, Fira is compatible with Unsloth by directly replacing the original optimizer. We will give it a try. If you have any other good suggestions, feel free to share with us.
Looks great and thank you!
Btw, I am interested in, for example, fine-tuning (SFT) llama 3.1 8B on a single RTX4090 (24GB) with math datasets. Is this supported by Fira? It seems that the paper does not talk about math or (complicated) reasoning, so I wonder it is because this approach does not work for math, or because it is simply not tested yet. (I believe it is the latter; but in case it is the former I ask this)
Thanks for your advice. Due to the DDL, we haven't evaluated our method on math reasoning datasets, and we will complete it soon.
Thanks for the information!
Hi thanks for the library! https://github.com/unslothai/unsloth is a library that supports fast PEFT fine-tuning. Therefore, I wonder whether this is / will be compatible with that?