Closed Yu-Fangxu closed 6 months ago
Hi, Fangxu!
Thanks for the question! Our experiments were conducted in a fully-parameter fine-tuning setting. In fact, 4 A100 80G GPUs are enough for our 7B experiments, including SFT and DPO. To implement LoRA in your training, you will need to modify fastchat/train/train.py
and fastchat/train/train_dpo.py
. Maybe you can see fastchat/train/train_lora.py
for the reference implementation of integrating LoRA with FastChat.
Hi, Thanks for your wonderful work, I noticed that you fine-tuned LLMs with 8 A100 GPUs. Have you ever tried training with LoRA for less consumption of computational resources? Thanks~