-
跑baseline运行evaluate函数中计算spice的部分产生如下错误:
```shell
PermissionError: [Errno 13] Permission denied: '/lustre/home/acct-stu/stu282/Tools/miniconda3/envs/py3.10.11/lib/python3.10/site-packages/pycocoevalc…
-
in the BLIP-2 paper,
"We propose Q-Former as the trainable module to bridge the
gap between a frozen image encoder and a frozen LLM. It
extracts a fixed number of output features from the image
en…
-
# The Illustrated Image Captioning using transformers - Ankur NLP Enthusiast
The Illustrated Image Captioning using transformers
[https://ankur3107.github.io/blogs/the-illustrated-image-captioning-u…
-
Hi, I used BlipForConditionalGeneration from transformers for image captioning.
I want to visualize the reason of generated caption (word by word) like GradCAM.
I found a code from Albef (https://…
-
https://arxiv.org/abs/1803.05526
https://arxiv.org/abs/1707.07998
https://arxiv.org/abs/1909.02201
https://arxiv.org/abs/1811.10787
-
Hi, I used BlipForConditionalGeneration from transformers for image captioning.
I want to visualize the reason of generated caption (word by word) like GradCAM.
I found a code from Albef (https://…
-
I want to use Beit3 using weight beit3_large_patch16_480_coco_captioning for image captioning on my custom images. I have download the weights and .spm file and using the following command:
!python -…
-
hi there,
is there any example (if possible) to use Llama2 for image captioning ?
thank you
-
List of images that will need additional attention during contextualization
**Lesson 1**
[Image](https://github.com/nasa/Transform-to-Open-Science/raw/v1.0.0/Module_5/images/media/image7.jpeg) w…
-
### Describe your use-case.
I recently was using OneTrainer UI to auto-caption ~750k images using WD14 tagger. I have a dual GPU machine, but it was only utilizing one of them (and was running at ver…