-
Hitting this issue when decoding: any thoughts?
```
File "/home/ubuntu/wbc/captioning/InternLM-XComposer/projects/ShareGPT4V/run_captioning.py", line 102, in gen_json
captions = eval_mode…
-
Hi all,
I have an issue regarding inputs to attribute method of integrated gradient algorithm. I am using the GIT model for image captioning and defined the forward function to return one token_id of…
-
Hello!
I have a question about extracting region features for image captioning:
- in VinVL paper, it states that 2048 region features are stacked with 6 positionally encoded features (bbox, its h…
-
Hi, glad to see and use this cool project, thanks you.
I have a question: if it possible to batch predictions on Image captioning task?
I see https://github.com/salesforce/BLIP/issues/48 but it's no…
-
Add the audio transcription using whisper and image captioning with gpt4-v functions and implement them in the `gradio ui` notebook
-
Hi, thanks for your amaizing work, i'm enjoy to use BLIP which demonstrate impressive results:)
Now i have a question: how can i fine tune BLIP for Image Captioning task on custom dataset?
My dat…
-
I'm curious what prompts DynamiCrafter responds to? I see some in the examples, but I was wondering if there was a resource that had more info on what type of prompts it responds best to?
-
`processor = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-base")`
`model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-base")`
Dataset[datatset3…
-
Hello, I wondered if there was a way to train the model and new data. I am sorry if I missed the documentation somewhere.
-
Hi,
I want to finetune the model on my own dataset. How should I prepare the stage1 and stage2 training data? What is the difference? The description of caption_stage1_train.tsv and caption_stage2_t…