microsoft / LLaVA-Med

Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabilities.
Other
1.59k stars 202 forks source link

Possible Access to LlaVa-Med using BioMed CLIP? #65

Open tj-zhu opened 7 months ago

tj-zhu commented 7 months ago

Thank you very much for sharing the data and model checkpoints in support of the community!

In the paper it mentioned

"the initialization of vision encoder from BioMed CLIP is slightly better than from general-domain CLIP".

I am wondering if it's possible to get access to the LlaVa-Med using BioMed CLIP?

Thank you again for the great work from the team.