-
I want to use Beit3 using weight beit3_large_patch16_480_coco_captioning for image captioning on my custom images. I have download the weights and .spm file and using the following command:
!python -…
-
https://arxiv.org/abs/1803.05526
https://arxiv.org/abs/1707.07998
https://arxiv.org/abs/1909.02201
https://arxiv.org/abs/1811.10787
-
Hi, I used BlipForConditionalGeneration from transformers for image captioning.
I want to visualize the reason of generated caption (word by word) like GradCAM.
I found a code from Albef (https://…
-
Hi, I used BlipForConditionalGeneration from transformers for image captioning.
I want to visualize the reason of generated caption (word by word) like GradCAM.
I found a code from Albef (https://…
-
Thank you for your work!
I see that music QA dataset is released on huggingface, but I'd like to know whether the generated music captioning dataset is also released.
-
### Describe your use-case.
I recently was using OneTrainer UI to auto-caption ~750k images using WD14 tagger. I have a dual GPU machine, but it was only utilizing one of them (and was running at ver…
-
hi there,
is there any example (if possible) to use Llama2 for image captioning ?
thank you
-
I think Ben's solution to captioning should work fine, ultimately, but some work might need to happen to get css to be respected by the pkgdown machinery as well as the rendering of the README.md file…
-
It be ideal if the Close Captioning/subs could be accessed on videos put onto the Internet Archive.
As is, even going in to interact mode, you can't turn them on.
Was tempted to put this as a b…
-
Warning: coco-caption not available
cider or coco-caption missing
Where is coco-captioning?
## Requirements
- cider (already been added as a submodule)
- coco-caption (already been added as…