salesforce / BLIP

PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
BSD 3-Clause "New" or "Revised" License
4.85k stars 648 forks source link

nocaps dataset #50

Closed yxoh closed 2 years ago

yxoh commented 2 years ago

can you share the URL to download the NoCaps dataset or the code to download nocap images? The official website only provides the URL for each image.

LiJunnan1992 commented 2 years ago

Hi, you can download from the URLs

woctezuma commented 2 years ago

See https://github.com/rom1504/img2dataset

lorenmt commented 2 years ago

Like to follow this issue again...

I am wondering what's the best tool to convert urls to images? I have downloaded the "ccs_synthetic_filtered_large.json" linked from this repo, while using requests.get() to download each url roughly resulted in 15% invalid urls across first 10k files I downloaded.

Is this normal? Or I should follow the package linked from @woctezuma , but if using that method, it would not easy to apply the filtered captions generated from BLIP right?

Really appreciate for the help and detailed instructions..

zjizhang commented 3 months ago

can you share the URL to download the NoCaps dataset or the code to download nocap images? The official website only provides the URL for each image.

I also want to know how to get NoCaps dataset