-
Hey, thanks for creating these notebooks! But I am trying to run Idefics_FT, and unfortunately, it isn't working... I run into an out of memory error when calling trainer.train() even though I am runn…
-
When I execute the following code, I cannot connect. But other models are connectable, what causes this?
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it…
-
I am reproducing the model on V100 GPU. If anyone is doing the same, I hope we can communicate and exchange ideas together. My wechat : Anymake_ren
1、Flickr 30k :
http://shannon.cs.illinois.edu/D…
-
Hi, thank you so much for the great work and releasing the code. I would like to study on videoQA ability of this model, specifically or MSVD-QA or TGIF-Frame, is that possible for us to download the …
-
[paper](https://arxiv.org/pdf/2310.03744.pdf)
see llava https://github.com/long8v/PTIR/issues/128#issue-1749571159 here
## TL;DR
- **I read this because.. :** aka LLaVA1.5 / ShareGPT4V에서 LL…
-
Why is OBELICS generally better than MINT-1T (HTML)? Is the main advantage of MINT-1T over OBELICS primarily related to PDFs?
-
[paper](https://arxiv.org/abs/2311.04257)
## TL;DR
- **I read this because.. :** very recent VLM model
- **task :** VLM + LLM
- **problem :** multi-modal task는 LLM freeze 시키고 사실상 V+L을 잘하려고…
-
Thanks for your great codes!
In your paper, running the pre-training experiments needs 64 V100 GPUs.
For research purposes, it is too heavy.
If using a small batch size, the performance would dr…
-
**Describe**
Model I am using (UniLM, MiniLM, LayoutLM ...): BEIT-3
I want to evaluate BEiT-3 finetuned model on VQAv2.
[https://github.com/microsoft/unilm/blob/master/beit3/get_started/get_start…
-
### Question
Hi @haotian-liu, thanks for your great project.
As mentioned in the paper, the statistics of llava_v1_5_mix665k.json is shown as Table 7:
That is, 158+40+83+72+9+80+50+22+30+86…