Closed FeiElysia closed 1 month ago
哥们你解决这个问题了吗
don't bother with the adapter, check the new training script, you can just start from the already fine-tuned llava-onevision model in --model_name_or_path
Thank you for your comment. Yes, you can run the script if you don't use --pretrain_mm_mlp_adapter
. If you still want to use your own adapter, you can set deepspeed
to scripts/zero2.json
. This can address the mismatching. There are some bugs when using lazy load with zero3.json
Dear authors,
Thanks for your promising work, I am trying to fine-tune LLaVA-OV on my own datasets, I modified the
finetune_onevision.sh
as follows:But I got the following error:
It seems that the projector is not initialized correctly. The code can run when I comment out
--pretrain_mm_mlp_adapter
. If there any special setting when loading projector?