-
### System Info
Use the ghcr.io/huggingface/tgi-gaudi:2.0.1 official docker image.
As shown in https://github.com/huggingface/tgi-gaudi/blob/habana-main/docs/source/supported_models.md?plain=1, l…
-
Hello. Thanks for your excellent work!
Earlier, I reproduced LLaVA-NeXT-Image training and got the desired performance, and I am now trying to reproduce LLaVA-NeXT-Interleave training. I would like…
-
I'm trying to instruction tune llava-next models following the llava_vsft.py examples shared for llava-1.5.
```
python vsft.py \
--dataset_name="HuggingFaceH4/llava-instruct-mix-vsft" \
--…
-
`bash playground/demo/interleave_demo.py --model_path path/to/ckpt`
The execute code should be run with python not bash.
And How can this code specify the input image sequence? It appears to be jus…
-
### System Info
- `transformers` version: 4.42.4
- Platform: Linux-6.2.0-39-generic-x86_64-with-glibc2.35
- Python version: 3.11.9
- Huggingface_hub version: 0.23.4
- Safetensors version: 0.4.3
…
-
### The model to consider.
LLaVA-NeXT-Video* (LlavaNextVideoForConditionalGeneration)
### The closest model vllm already supports.
Llava-Next (LlavaNextForConditionalGeneration)
### What's your …
-
Thanks for your great work! I'm wondering if u can share the loss curve for training llava-next-llama3? I've observed some different behaviour compared to training llava-next-vicuna-7b. I'm wondering …
-
Hi, I notice llava-next has published new version of llava-next-video model with llava-qwen and siglip vision tower, I wonder do have plan to support siglip in sglang? Thanks~
-
### The model to consider.
The llava-next-video project has already been released, and the test results are quite good. Are there any plans to support this project?
`https://github.com/LLaVA-VL/LLaV…
-
Great work! I notice the LLaVA-NeXT-Qwen2 (image model) can achieve a surprising 49.5 Video-MME results. In contrast, the LLaVA-NeXT-Video (Llama3) can only achieve a 30+ Video-MME score (according to…