-
### What happened?
Hi there.
I got an unexpected slot_id and responses when sending 4 concurrent requests to a llama-server started with:
```bash
./llama.cpp-b3938/build_gpu/bin/llama-server -…
-
### What is the issue?
Currently Ollama can [import GGUF files](https://github.com/ollama/ollama/blob/main/docs/import.md). However, larger models are sometimes split into separate files. Ollama shou…
-
Installation went fine but I get the following error when trying to invoke the assistant:
`Sorry, there was a problem talking to the backend: RuntimeError('llama_decode returned 1')`
![image](ht…
-
Hey I'm trying to install ggml vicuna following this link: https://agi-sphere.com/install-textgen-webui-mac/
The texgen-webui works as I tested it on a small LLM model, however, it doesn't work for…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
It's my first bug report on the GitHub, please, don't be mad if I've done something wrong.
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [☑…
-
this is what I get
meriem@Home:~/llama-gpt$ ./run.sh --model 7b --with-cuda
[+] Building 4.2s (30/30) FINISHED …
-
**Is your feature request related to a problem? Please describe.**
When generating chat completion, it is hard-coded to generate a non-standard prompt template that looks something like:
```
### …
-
See https://github.com/abetlen/llama-cpp-python/issues/223#issuecomment-1556207716
A bug looks to be with `libllama`, but in order to log the bug with [llama.cpp](https://github.com/ggerganov/llama…
-
**Describe the bug**
After following the installation instructions for Mac (Apple Metal) on the 'Getting Started' I try to do SDG and see this:
```
$ ilab data generate
INFO 2024-08-22 13:45:45,…