-
Hi,
Can you add the VQA fine-tuning function of BLIP2?
In the paper, when you fine-tune the VQA task, you will fine-tune the image encoder. When I use the `freeze_vit: False` command.
But I encoun…
zhl98 updated
4 months ago
-
## タイトル: シーングラフを用いた目標指向セマンティック通信による無線画像質問応答
## リンク: https://arxiv.org/abs/2411.02452
## 概要:
通信および計算能力への需要が高まるにつれ、従来のビット指向通信は、特にミッションクリティカルで計算量の多いアプリケーションにおいて、これらの厳しい要件を満たせなくなってきています。代表的なアプリケーションであ…
-
![image](https://github.com/user-attachments/assets/75ef8603-9cba-404c-9435-7a946e9bb2b0)
-
Hi thanks for your work and public release of the code.
I have checked your code and I could not find the generate function of your model while using the VQA model. I want to be able to input new q…
-
Hi Oscar teams:
I read your code and find that some parser_argument may loss, like "--model_name_or_path vinvl/model_ckpts/vqa/base/checkpoint-2000000" and "--tokenizer_name". Can you provide this fi…
-
Hi we are working on finetuning VQA with BLIP2. Any instructions on how to modify the codes? When will the finetuning codes be released?
-
https://paperswithcode.com/dataset/dvqa
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
```
- `llamafactory` version: 0.9.1.dev0
- Platform: Linux-4.19.90-2107.6.0.0192.8.oe1.bclinux.x86_64-x…
-
Hi, thank you for your excellent work! I am interested in your P-VQA dataset. Would you please explain the numbers of the label feature in the pvqa dataset? Can the label be transferred back to natura…
-
Dear authors,
Thanks for the great work! I wonder to know the zero-shot performance of InstructBLIP on OK-VQA Dataset. However, it's not report in the paper. I reproduced this and got 57.6%. Can yo…