-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [x] 2. The bug has not been fixed in the latest version.
### Describe the bug
Hi 👋, I quantized llava 1.5 1…
-
### System Info
TGI 2.0.3 Docker
2 A10G's and 1 A100 without `num-shard`
### Information
- [X] Docker
- [ ] The CLI directly
### Tasks
- [X] An officially supported command
- [ ] My own modifica…
-
Will there ever be a version with Command-R (https://huggingface.co/CohereForAI/c4ai-command-r-v01) support?
llama.cpp supports this model for a long time.
-
I just ran this code
```python
from llama_cpp import Llama
from llama_cpp.llama_chat_format import Llava15ChatHandler
# Create a chat handler
chat_handler = Llava15ChatHandler(clip_model_path="…
-
The [server](https://github.com/ggerganov/llama.cpp/tree/master/examples/server) example has been growing in functionality and unfortunately I feel it is not very stable at the moment and there are so…
-
mlx 0.13.1
mlx-lm 0.13.1
mlx-vlm 0.0.5
```
import mlx.core as mx
from mlx_vlm import load, generate
model_path = "ml…
-
### System Info
- `transformers` version: 4.40.0.dev0
- Platform: Linux-5.19.0-051900rc6-generic-x86_64-with-glibc2.35
- Python version: 3.9.18
- Huggingface_hub version: 0.21.1
- Safetensors ver…
-
Thank you for this wonderful work, however when i tried to run the inference demo i got `run_llava.py: error: unrecognized arguments: --model-name Efficient-Large-Model/VILA-13B`
so i changed `--mo…
-
Hi authors,
Thanks for the great work!
However, I cannot reproduce the numbers reported in the paper using your code. I use the **LLaVA-1.6-vicuna-7B** model.
Open-ended QA
| | MSVD-QA | MSRVT…
-
[Here](https://github.com/opendatalab/HA-DPO/blob/42f72c536984c6ded016e89b70266f29f2428f33/ha_dpo/models/llava-v1_5/train_dpo.py#L218) Why does this variable A need to be multiplied by two here?
Be…
tbbbk updated
4 months ago