-
I get this error when i Try to run a query
Truncation was not explicitly activated but `max_length` is provided a specific value, please use `truncation=True` to explicitly truncate examples to max…
-
~/llama-node/packages/llama-cpp$ node example/mycode.ts
llama.cpp: loading model from /llama-node/packages/llama-cpp/ggml-vic7b-uncensored-q5_1.bin
llama_model_load_internal: format = ggjt v2 (…
-
Can someone help me configure this
Using Python 3.11
ROCm Version 5.5.1
× Building wheel for llama-cpp-python (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [55 lines o…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
# Feature Description
Please provide a detailed written description of what you were trying to do, and what you expected `llama.cpp` to do as an enhancement.
# Motivation
It sounds like it's …
-
Create a struct `ggml_metal_locals` and populate using `GGML_TENSOR_LOCALS` similar to what we do in `ggml.c`:
https://github.com/ggerganov/llama.cpp/blob/3b4bab6a38502d9e68587c2c19f26472480ec4dd/g…
-
I am not able to find much on batching support. But it appears that the downstream llama.cpp supports it.
https://github.com/ggerganov/llama.cpp/issues/4372
Any plans to expose this feature in k…
sirmo updated
5 months ago
-
The current state of the testing framework is pretty bad - we have a few simple test tools in [tests](https://github.com/ggerganov/ggml/tree/master/tests), but these are not maintained properly and ar…