magic-research / PLLaVA

Official repository for the paper PLLaVA
578 stars 40 forks source link

How to fintune base on your fintuned result #56

Open liuao743 opened 5 months ago

liuao743 commented 5 months ago

You fine-tuned the language model, etc., using lora fine-tuning on the basis of the original model llava-hf/llava-v1.6-vicuna-7b-hf, but your open source weights (ermu2001/pllava-7b) seem to contain only lora results. Reason: I used ermu2001/pllava-7b and ermu2001/pllava-13b as repo_id parameters to train, and their loss decreased from the order of 10.

If I use llava-hf/llava-v1.6-vicuna-7b-hf as the parameter of repo_id, although loss is normal, it is equivalent to that I have not used the weight of your lora.

After checking the fine-tuning code, I found that only the parameter repo_id was used to pass the model path. After further checking the training code, I did not find any place where lora weight could be passed. May I ask how can I continue to fine-tune the model based on your fine-tuning?

liuao743 commented 5 months ago

@ermu2001

gaowei724 commented 4 months ago

Hi, I think the solution to your problem is setting repo_id=llava-hf/llava-v1.6-vicuna-7b-hf and pretrained_path=ermu2001/pllava-7b, reference to https://github.com/magic-research/PLLaVA/issues/45