salesforce / BLIP

PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
BSD 3-Clause "New" or "Revised" License
4.8k stars 641 forks source link

Do you try continuous bootstrapping? #89

Open TheShadow29 opened 2 years ago

TheShadow29 commented 2 years ago

Hello, thanks for open-sourcing your code, this is a very inspiring work. I had a question: do you perform or experiment with bootstrapping continuously? Like currently it is train on pre-trained dataset, fine-tune on coco-retrieval, captioning, and then bootstrap to get new pre-train dataset. Now, pre-train on this and repeat the bootstrap process again?

How expensive is this operation?

LiJunnan1992 commented 2 years ago

Hi, thanks for your question. Multi-round bootstrapping can be expensive, thus we have not tried it out yet. We do expect that continuous bootstrapping would give further performance improvement.