camenduru / LLaVA-colab

208 stars 31 forks source link

🐣 Please follow me for new updates https://twitter.com/camenduru
πŸ”₯ Please join our discord server https://discord.gg/k5BwmmvJJU
πŸ₯³ Please join my patreon community https://patreon.com/camenduru

πŸ¦’ Colab

Colab Info
Open In Colab πŸŒ‹ LLaVA_13b_4bit_colab 13B (4bit)
Open In Colab πŸŒ‹LLaVA_13b_4bit_vanilla_colab 13B (4bit) (without gradio)
Open In Colab πŸŒ‹LLaVA_13b_4bit_caption_colab 13B (4bit) (without gradio - image caption loop)
Open In Colab πŸŒ‹ LLaVA_7b_8bit_colab 7B (8bit)
Open In Colab πŸŒ‹ LLaVA_7b_colab 7B (16bit) (Pro High-RAM 😐 22GB RAM 14GB VRAM)

⚠ Tutorial

https://www.youtube.com/watch?v=o7zQAa0NPds

After clicking the third cell, please wait for the model to load.
~14.7GB for 16bit ~8GB for 8bit ~5 minutes
We will not receive any output because it will be running in another thread.

Screenshot 2023-10-09 210109a

Main Repo

https://github.com/haotian-liu/LLaVA

Paper

https://arxiv.org/abs/2304.08485

Page

https://llava-vl.github.io/

Output

Screenshot 2023-10-07 115057

Screenshot 2023-10-14 164638