-
- [LLaVA-Med: Training a Large Language-and-Vision Assistant for Biomedicine in One Day](https://arxiv.org/abs/2306.00890)
- [MEDITRON-70B: Scaling Medical Pretraining for Large Language Models](http…
-
Hi,
Thanks for sharing the model and code with us.
I am trying to using Vision Foundation Model for a zero shot classification problem.
It is possible with **OpenGVLab/InternVL-14B-224px** bu…
-
https://huggingface.co/blog/vision_language_pretraining
-
- [ ] [DeepSeek-VL: Towards Real-World Vision-Language Understanding](https://arxiv.org/html/2403.05525v2)
# DeepSeek-VL: Towards Real-World Vision-Language Understanding
**Abstract**
We present De…
-
Pose a question about one of the following articles:
“[Online images amplify gender bias](https://www.nature.com/articles/s41586-024-07068-x),” 2024. Guilbeault, Douglas, Solène Delecourt, Tasker …
-
-
- [ ] [Title: "Yi Model Family: Powerful Multi-Dimensional Language and Multimodal Models"](https://arxiv.org/html/2403.04652v1)
# Title: "Yi Model Family: Powerful Multi-Dimensional Language and Mul…
-
Hi, the paper "Image as a Foreign Language: BEIT Pretraining for All Vision and Vision-Language Tasks" is really interesting, and the results are astonishing well, congratulations!
I am writing ju…
-
Hi, did u first train the projector, and then train projector + LLM, what's the detail of them.
-
Dear Developers:
Thank you to the BAAI team for open-sourcing the Bunny model. I've been actively exploring it these past few days. I have a few doubts regarding the deployment of the model, and I …