-
### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What would your feature do ?
see this discussion: https://github.com/Mik…
-
Would love to see: [Bunny-Llama-3-8B-V](https://huggingface.co/BAAI/Bunny-Llama-3-8B-V) included in the Ollama models.
> Bunny is a family of lightweight but powerful multimodal models. It offers m…
-
Sorry to bother, but in the usage guidance in eva-clip, there is a line: "from eva_clip import create_model_and_transforms, get_tokenizer".
However, I could not find such a file in the folder.
I un…
-
Traceback (most recent call last):
File "/cpfs/user/mingcanma/workspace/code/openseg/APE-main/demo/demo_lazy.py", line 135, in
demo = VisualizationDemo(cfg, args=args)
File "/cpfs/user/min…
-
不下载模型,
settings in ComfyUI.
Load model: EVA01-g-14/laion400m_s11b_b41k
Loading caption model blip-large...
Loading CLIP model EVA01-g-14/laion400m_s11b_b41k...
Loaded EVA01-g-14 model config.
Un…
-
Here was the error when using powershell, creating venv, following instruction and installing the requirements.txt, and running python app.py
```
C:\users\newpc\downloads\pullid\pulid\venv\lib\sit…
-
We know that MiniGPT-v2 consists of Visual backbone (ViT), Linear projection layer (Linear), and Large language model (LLaMA2), where Visual backbone uses the ViT of eva_clip_g. However, vis_processor…
-
I find that the eva-clip model has an extra inner_attn_ln layer compared to the original pretrained model.
[EVA-CLIP](https://github.com/baaivision/EVA/blob/master/EVA-CLIP/rei/eva_clip/eva_vit_mo…
-
First thanks for your great job!
Now We're trying to replace the vision encoder in llava, i.e., clip-l-336, with RADIO. Under the default LLaVA 1.5 settings, we pretrain a multimodal projection MLP a…
-
I follow this tip:
```
Python 3.10.8 (main, Nov 24 2022, 14:13:03) [GCC 11.2.0]
Version: v1.9.3
Commit hash: 1c0a0c4c26f78c32095ebc7f8af82f5c04fca8c0
Launching Web UI with arguments: --port=6006 …