-
Additionally, it might be interesting to consider adding GGUFed version of more models?
![image](https://github.com/user-attachments/assets/0414d354-47db-4dfa-b4de-cd3443217bbe)
-
### Summary
A segment fault is raised while doing two consecutive `image-edit` requests. In the test, the first request is done successfully. Then, the second request is sent; and during the runnin…
-
**Describe the bug**
I have models on my system that I'd already downloaded in /srv/models.... I didn't realize there were three model paths in the config.yaml to start so I got a traceback when i …
-
Raising this issue to help integrate **Llama Guard 3-11B-vision Model Card** to detect harmful multimodal prompts and text responses to these prompts and safeguard content for both LLM inputs (prompt …
-
# 🎉 Open Call for Contributions to the LLaMA Recipes Repository
Hey there! 👋
We are excited to open up our repository for open-source contributions and can't wait to see what recipes you come up…
-
In https://github.com/instructlab/instructlab/pull/1797 we enabled batching unconditionally with remote endpoints.
However, we know it doesn't work with llama-cpp - see e.g. https://github.com/inst…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
Please create the following browser wasm demos-
1) Stable diffusion with W8A8 quantization- This is important because the stable diffusion [demo](https://intel.github.io/web-ai-showcase/) which I sa…
-
The llama.cpp integration within the playbook does not works, anyway i have manually created the gguf file but when i try to serve the model using the llama.cpp server i am getting the following error…