Open GMBarra opened 1 year ago
It is suggested to initialize from stage-2 pretrained model for finetuning on captioning dataset. You can create a new dataset builder following the instructions, and then modify caption_coco_ft.yaml to use your own dataset.
Hi, I am interested in fine-tuning the BLIP2 model on a custom dataset for captioning or classification tasks. My custom dataset is formatted similarly to the COCO dataset, consisting of a dictionary with image paths and corresponding image captions. I have two questions regarding this:
caption_coco_ft.yaml
file to redirect the path to my custom dataset, or do I need to create an entirely new dataset format?Any help or guidance you could provide would be greatly appreciated.