issues
search
salesforce
/
BLIP
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
BSD 3-Clause "New" or "Revised" License
4.85k
stars
648
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Access Denied for answer list for the test dataset for vqa dataset
#224
amitpadegal
opened
7 hours ago
0
run demo on colab bug
#223
zkailinzhang
opened
3 days ago
0
Repetitive Output for All images
#222
bruceMug
closed
1 week ago
0
What key design processes make BLIP have no limitations on text input length compared to CLIP?
#221
Yang-bug-star
opened
1 month ago
0
How can I do inference with pretrained models rather than finetuned models
#220
Yang-bug-star
opened
1 month ago
0
What does the model(**inputs, output_attentions=True) output?
#219
ZhanliangAaronWang
opened
1 month ago
0
FileNotFoundError: [Errno 2] No such file or directory: 'export/share/datasets/vision/coco/images/val2014/COCO_val2014_000000184613.jpg'
#218
Jingut
opened
2 months ago
1
How to train itm from itc
#217
Raion-Shin
opened
3 months ago
0
is it possible to change the text LLM of blip to a different LLM?
#216
KAABSHAHID
opened
3 months ago
0
Caption on ImageNet-Dogs
#214
LouisDong95
closed
1 week ago
0
fix the wrong words
#213
ArtificialZeng
opened
3 months ago
2
knowledge distillation
#212
sssssshf
opened
4 months ago
0
blip_vqa error
#211
AIWASS23
opened
4 months ago
2
fixed runtime device error
#210
antelopoul
opened
6 months ago
0
How to use roberta as the decoder
#209
xiweideng
opened
7 months ago
0
The rank_answer function in BLIP is different from that in ALBEF
#208
littleFlyDance
opened
7 months ago
0
Question or bug in blip_pretrain.py
#207
LiGuo12
opened
7 months ago
0
How to retrive the raw attention scores or logits from blip model ( image captioning)
#206
umme17
opened
7 months ago
0
I want to use the existing image-text pedestrian dataset and finetune the BLIP model. Should I use pre-trained checkpoints weights or finetuned checkpoints weights?
#205
shams2023
opened
8 months ago
0
Image-Text Retrieval
#204
mjjc111
opened
8 months ago
0
LAION 115M dataset has 11164.tar?
#203
jacob-kang
opened
8 months ago
0
Error while running Colab demo
#202
staru09
opened
9 months ago
1
stable-diffusion RuntimeError: Couldn't fetch BLIP.
#201
saiheitor
opened
9 months ago
1
ITM Loss Stuck at 0.63
#200
bfan1256
opened
9 months ago
3
How to use the retrival large model for image-text prediction (model_large_retrieval_coco) ?
#199
caydenwei
opened
9 months ago
0
Blip Replicate Interface Is Down
#198
hashnimo
closed
9 months ago
1
ModuleNotFoundError: No module named 'ruamel_yaml'
#197
WenjunHuang94
opened
10 months ago
2
web demo issue
#196
hhzhao0525
opened
10 months ago
0
BlipForImageTextRetrieval loss returned when labels is provided
#195
AIIRLab
opened
10 months ago
0
RuntimeError: Default process group has not been initialized, please make sure to call init_process_group (train_caption.py)
#194
Y-HuiMing-Y
opened
10 months ago
1
what's the effect of 'image_queue' and 'text_queue'
#193
FengWu-PKU
opened
11 months ago
1
How does the BLIP model use half-precision (FP16) inference?
#192
WKaiH123
opened
1 year ago
0
About the ViT of BLIP
#191
LWShowTime
opened
1 year ago
0
Need clearly Understand of each checkpoint
#190
p1k0pan
opened
1 year ago
0
I am having trouble running evaluation code
#189
jyrana
closed
1 year ago
5
相似图像生成的caption一样,该如何解决?
#188
shams2023
opened
1 year ago
0
Video subtitle generation
#187
Levi-arch1
opened
1 year ago
0
一张3090卡去微调COCO检索,需要多长时间?
#186
shams2023
opened
1 year ago
2
Add auto-label guide
#185
capjamesg
closed
2 months ago
2
New ViT findings via registers (2309.16588)
#184
Infinitay
opened
1 year ago
0
The pre-trained BLIP model is used directly to perform caption operation, but the generated caption effect is not good
#183
shams2023
closed
1 year ago
0
.
#182
shams2023
closed
1 year ago
0
No scores of VQA evaluation
#181
p1k0pan
closed
1 year ago
5
This error indicates that your module has parameters that were not used in producing loss
#180
ericosmic
opened
1 year ago
0
Request for BLIP Pretrain Training Logs and Dataset Inquiry
#179
Aitical
opened
1 year ago
0
what is mean of 'question_states += [question_output.last_hidden_state[b]]*n'
#178
ericosmic
opened
1 year ago
0
retrieve output not fix
#177
ltm920716
opened
1 year ago
0
ERROR: Could not build wheels for tokenizers, which is required to install pyproject.toml-based projects
#176
TheOneTrueGuy
opened
1 year ago
4
Can BLIP generate more words image caption?
#175
uestcMeng
opened
1 year ago
7
Cosine between image_features and text_features taken from BLIP_Extractor_Features gives bad results
#174
aTunass
opened
1 year ago
0
Next