-
### Checklist
- [ ] 1. I have searched related issues but cannot get the expected help.
- [ ] 2. The bug has not been fixed in the latest version.
- [ ] 3. Please note that if the bug-related issue y…
-
![5D190FA6EE718064BEC8DBD812DCF1B3](https://github.com/user-attachments/assets/f7fd1920-6046-46e7-9162-f6b30ee15a8e)
I downloaded siglip-so400m-patch14-384 and write down the path. What else do I n…
-
很棒的工作!!!
但是有一个问题我有点疑惑,我发现在train_llava15.py文件中vision_tower似乎是带有梯度的,这样训练出来的模型视觉模块的参数应该是训练过的,但是在我下载的huggingface上rlaif-v 7b的参数到本地后运行chat.py文件会报以下的警告,似乎是视觉模块的参数没有被加载进来,并且还需要再从huggingface上下载clip-vit的参数:
…
-
### Problem
Hello,
I'm very pleased to see exllama getting vision capabilities for the first time with Pixtral!
You hinted at supporting new models in the release notes. What models are you h…
-
As title says. I can upload images to llava but not llama3.2-vision
llama3.2-vision:11b-instruct-q8_0
Nayar updated
2 weeks ago
-
### What is the issue?
I tried to import finetuned llama-3.2-11b-vision, but I got "Error: unsupported architecture."
In order to make sure my model is not the problem, I downloaded [meta-llama/Ll…
-
i saw https://huggingface.co/Vision-CAIR/LongVU_Llama3_2_1B exists .
Is it image or video part ?
could it be combined with LongVU_Llama3_2_3B ? (image or video) and what
hardware requirements ?
-
Hi, thanks for your work.
When I run the demo code from: https://huggingface.co/lmms-lab/LLaVA-Video-72B-Qwen2 in your LLaVA-NeXT repository, some errors happened:
```
size mismatch for vision_mode…
-
reported by RauelON
-
Hi,
I am trying to load a Phi-3.5-3.8B-vision-instruct-Q8_0 GGUF model using the command for loading local GGUF file:
./mistralrs-server -i gguf --quantized-model-id path/to/files --quantized-f…