THUDM / CogVLM2

GPT4V-level open-source multi-modal model based on Llama3-8B
Apache License 2.0
1.42k stars 77 forks source link

Finetuning learning rate is very small #94

Open orrzohar opened 2 weeks ago

orrzohar commented 2 weeks ago

System Info / 系統信息

Hi

I noticed that the learning rate for the instruction tuning phase (1e-7) is 100 times smaller than what was reported in the COG-VLM technical report (1e-5).

https://github.com/THUDM/CogVLM2/blob/57e5a80e996a0e36d9302e9efa3f63cfc29d3368/finetune_demo/peft_lora.py#L185

What is the reason for this? is this due to the LLaMA3? why is COG-VLM2 so much less stable?

++ When you fine-tuned COG-VLM2, did you also only do LoRA?

Best, Orr

Who can help? / 谁可以帮助到您?

No response

Information / 问题信息

Reproduction / 复现过程

-

Expected behavior / 期待表现

-