Open Harry-mic opened 11 months ago
Hi, it's unlike the problem of the difference between 7b and 70b.
What is the version of the PEFT you used? We use peft==4.0.0
in our experiments, and perhaps the behavior is different in newer or older versions.
Thanks for your reply!
I use the peft==0.6.3.dev0
version and I think that's the point.
Hi! I encounter an issue that when doing the Step3(SFT).
The function "get_accelerate_model" in qlora_model.py sets the adapter_name="lora_default". This results in an error that the trainable parameters are set to 0.0 rather than 1.6% of the full parameters:
I fix this by setting the adapter_name="default". I am finetuning a llama-2-7b-hf model and I wonder if it is a bug or an issue caused by the different finetuned model(7b and 70b)