BAAI-DCAI / Bunny

A family of lightweight multimodal models.
Apache License 2.0
799 stars 61 forks source link

The accuracy of model test is poor #78

Closed aoji0606 closed 1 month ago

aoji0606 commented 2 months ago

Thanks for your work.

I use the siglip and llama3 for pretrain and fine-tune,but the test results were poor,here are the results: image Why is that?

Isaachhh commented 2 months ago

We use a better strategy to train Phi-3-Mini-based and Llama-3-8B-based Bunny, which would be open-sourced soon.

BTW, do you use Bunny-pretrain-laion-2m and Bunny-695k? It seems that your results are a little bit worse than expectation.

aoji0606 commented 1 month ago

Yes, I use Bunny-pretrain-laion-2m for pre-train, and Bunny-695k for fine-tune

Isaachhh commented 1 month ago

What's your pre-training learning rate?

aoji0606 commented 1 month ago

1e-3 for pre-train 2e-5 for fine-tune

Isaachhh commented 1 month ago

Do you use lora or full fine-tuning?

aoji0606 commented 1 month ago

I use full fine-tuning

Isaachhh commented 1 month ago

LoRA would convey a better performance.

aoji0606 commented 1 month ago

ahhh, ok, thank you

aoji0606 commented 1 month ago

hi~ What the learning rate should be set when using lora to fine-tune the llama3 model

Isaachhh commented 1 month ago

As default, 2e-4 and 2e-5.

aoji0606 commented 1 month ago

ok thank u

Isaachhh commented 1 month ago

Close the issue for now if there's no further discussions. Feel free to reopen it if there's any other questions.