-
I'm seeing a weird behavior with vision models.
I am using the Default LM Studio Windows config, which is the only one I have been able to get vision models to work with.
I have tried 2 differ…
-
Would love to see: [Bunny-Llama-3-8B-V](https://huggingface.co/BAAI/Bunny-Llama-3-8B-V) included in the Ollama models.
> Bunny is a family of lightweight but powerful multimodal models. It offers m…
-
### System Info
- `transformers` version: 4.45.0.dev0
- Platform: Linux-5.15.0-119-generic-x86_64-with-glibc2.35
- Python version: 3.10.12
- Huggingface_hub version: 0.24.6
- Safetensors versio…
-
I use GPT-4o is running ok.
But when I changed to the local model, I used some error message.
EXCEPTION: 'function' object has no attribute 'name'
![image](https://github.com/onuratakan/gpt-compute…
-
**Describe the bug**
What the bug is, and how to reproduce, better with screenshots(描述bug以及复现过程,最好有截图)
I encountered an OOM error when trying to DPO MiniCPM-LLaMA-v-2.5 with my own dataset and 4 r…
-
To aid in the design for both of these:
- #331
- #556
I'm going to gather a bunch of examples of how different LLMs accept multi-modal inputs. I'm particularly interested in the following:
- …
-
### Describe the bug
When I toggle the option for multimodal the software crashes.
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
Turning on …
-
Hi, @lucidrains !
There was a promising research published this month (vs. RoPE-mixed (#25) in March), the so-called LieRE positional encodings generalize the kv-vector rotation to any numbers of d…
-
# Interesting papers
- Yan 2024 - An Object is Worth 64x64 Pixels: Generating 3D Object via Image Diffusion [링크](https://omages.github.io/)
- Diffusion을 통해서 64 x 64 크기의 '부품 이미지' (Object image)…
-
When i trained llava-llama3 use your code, the log print tokenization mismatch as below.
how to fix it?
thanks!
WARNING: tokenization mismatch: 55 vs. 54. (ignored)
WARNING: tokenization m…