-
Thank you for your great work.
I want to evaluate llava-next 34b, but I only have A100 40B which is not enough for inference 34b in only 1gpu.
Can you give me some advices?
-
So,This repo help me a lot I even read your blog and this part
* Get pretrained vision encoder from SigLIP or CLIP (both come in difference sizes). Freeze weights (i.e. don’t update during
bac…
-
The [server](https://github.com/ggerganov/llama.cpp/tree/master/examples/server) example has been growing in functionality and unfortunately I feel it is not very stable at the moment and there are so…
-
I have a Qualcomm SnapDragon 8 Gen 3 equipped Android phone and trying run QNN SDK(/opt/qcom/aistack/qairt/2.23.0.240531) on this Android phone, I found there are two strange questions:
1.failed to…
-
Hey there!
First off: Thank you for this amazing model and all the work you put into it. So far this thing is really impressive (especially at this size).
I was messing around with it a bunch fo…
-
This issue to track supporting ollama as an alternative to openai.
-
```
export CUDA_VISIBLE_DEVICES="0,1,2,3"
python -m sglang.launch_server --model-path lmms-lab/llava-next-72b --tokenizer-path lmms-lab/llavanext-qwen-tokenizer --port=30010 --host="0.0.0.0" --tp-si…
-
### Reminder
- [X] I have searched the Github Discussion and issues and have not found anything similar to this.
### Motivation
The low footprint of Yi-VL's video memory and the high speed of its …
-
Thanks for the project ❤️ I made a colab. 🥳 I hope you like it. https://github.com/camenduru/LLaVA-colab
-
Dear author,
Thanks for releasing the amazing code. I'm trying to train the model using A100 (40G).
I loaded the pre-trained `mm_projector.bin` and run the `finetune.sh` script with video data.
…