-
Hello!
I'm new to the MLX ecosystem, and I came across the fact that there is a valid CLIP implementation available in the repository. Keeping in mind that [SigLIP](https://arxiv.org/abs/2303.15343…
-
### Feature request
Add support for export SigLIP models
### Motivation
As used by many SOTA VLMs, SigLIP is gaining traction and supporting it can be the step 1 to supporting many VLMs.
### Your …
-
首先感谢您的精彩工作。
目前我正在基于tinyllava模型利用tinychart数据复现训练流程,但是我发现bczhou/TinyLLaVA-3.1B-SigLIP中的visual encoder的image_size是384,vit_add_tome.py会将config中的image_size改成768。
因此在模型初始化时会基于image_size=768初始化sigLIP的p…
-
I am trying to add Mantis to the supported model list in VLLM or Sglang
-
Could you provide the result about DCI on LLaVA1.5 with clip-L, Vicuna-13B and 0.5M+0.6M?I can only find the 7B and siglip version.
-
model, image_processor = create_model_from_pretrained(model_name='ViT-SO400M-14-SigLIP-384', pretrained=self.image_tower_name,
image_mean=([0.5,0.…
-
Greetings.
Would like to ask two questions about TinyLLaVA-0.55B (With OpenELM-270M-Instruct):
1) From config.json provided in TinyLLaVA 0.55B's hf repo, seems it is using OpenELM-450M
2) The repo …
-
Hi~,I am recently trying to use the llava_onevision model, I try to follow the onevision tutorial, which seems pretty easy. I run the program exactly as the tutorial, the model is 0.5b_si. However, a …
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.…
-
No output text in comfui or text file after running. Logs show
/root/ComfyUI/models/clip/siglip-so400m-patch14-384
/root/ComfyUI/models/LLM/Meta-Llama-3.1-8B-bnb-4bit
Unused kwargs: ['_load_in_4b…