Closed guy-singer closed 9 months ago
Check out this pull request on
See visual diffs & provide feedback on Jupyter Notebooks.
Powered by ReviewNB
Hi @guy-singer thanks for adding a notebook to showcase the captions functionality. The notebook looks good as it is. I have some comments below. It would be great if we had these in the notebook. If not, then it's okay too :)
Typically in a notebook we will include a Download Dataset
section where we provide a tiny dataset to minimally reproduce the results in the notebook. This lowers the barrier for users to run this notebook without them having to think of a dataset to use. From a user experience perspective, this increases the chance of them running the notebook. I'd typically drop in a sentence or two to encourage users to run it on their own dataset.
There are readily available datasets in our examples like Oxford Pets, Imagenette, Mini COCO, etc. But if these are too large, then feel free to host a tiny dataset on cloud storage.
Also I'd add a sentence at the top of the notebook to tell users why they need the captioning capability in their workflow. Highlight benefits or include a motivation for the problem at hand.
I feel like we can unlock more capabilities of the VQA model by letting users specify their prompts. Currently, we are only prompting the model to identify indoor/outdoor which only applies to a niche problem.
Thanks Dickson, I have rewritten the notebook using the mini coco dataset, and added the description at the top as you have suggested. I will commit an updated notebook shortly.
I feel like we can unlock more capabilities of the VQA model by letting users specify their prompts. Currently, we are only prompting the model to identify indoor/outdoor which only applies to a niche problem.
Added capability to let users specify their VQA prompt. committing new notebook shortly with VQA example. Thanks Dickson.
all fixes and updates have been made, according to suggestions from @amiralush and @dnth
Notebook edited with clustering example as per @amiralush 's request
@guy-singer Not sure if I'm missing something, but why are we zero-ing the distance scores? Since it's an outlier report wouldn't it be useful to know if the image is far apart from all other images in the dataset?
@guy are you ready to deploy this?
@guy are you ready to deploy this?
@amiralush yes, it is ready and working well
These changes introduce the caption() function into the fastdup controller, allowing the user to caption their entire dataset or a subset of the dataset.
Additionally, an example notebook is added into the examples directory, demonstrating to the user how to conduct captioning.