Open cloneofsimo opened 6 months ago
Hey! Copying my answer from a previous issue - "The data - this is a big one - the full GRIT data might contain a lot of low quality images and/or prompts. Most of the data I used was either synthetic or filtered by CLIP aesthetic score. Try the mj_latents.npy and mj_text_emb.npy from here https://huggingface.co/apapiu/small_ldt/tree/main - this is higher quality synthetic data - I think about 600k examples if I remember correctly." Or you can use the data processing to download any data that has image and caption pairs from huggingface.
Hi there! Im trying to make minRF, and there was a pointer to here, I was wondering what dataset you used for this! Thanks!