Dear author:
I noticed you finetunning the whole LLM model without using LoRA. I wonder have you did some experiments on using or not using LoRA? Thank you.
Hi, because we donot find LoRA can significantly save memory or improve efficiency and performance with flash attention added. In this case, we maintain the full finetuning.
Dear author: I noticed you finetunning the whole LLM model without using LoRA. I wonder have you did some experiments on using or not using LoRA? Thank you.