-
I upgraded from an older version, and experienced a disturbingly long read-ahead time.
The load on my machine is about the same (a bit higher with python, but that's understandable)
I tried to speci…
-
### System Info
CPU x86_64
GPU L40s
TensorRT branch: main
commid id:b57221b764bc579cbb2490154916a871f620e2c4
CUDA:
| NVIDIA-SMI 535.154.05 Driver Version: 535.154.05 CUDA V…
-
# Expected Behavior
The llama-python-cpp should update and build.
# Current Behavior
I cannot build except 1.59 (which I just tried due to a few suggestions from a similar apparent bug in 1.60)…
-
When running the `python -m private_gpt` command I get the below error
```
ImportError: Could not import llama_cpp library.Please install llama_cpp with `pip install llama-cpp-python`.See the full…
-
# Prerequisites
I am running the latest code. Development is very rapid so there are no tagged versions as of now.
I carefully followed the [README.md](https://github.com/abetlen/llama-cpp-python/b…
DDXDB updated
7 months ago
-
I wanted to try a self-hosted offline LLM to summarize some notes and this looked like the most straightforward thing to try! I've attached a screenshot with hopefully relevant information
Tutorial…
-
Hi,
I am getting following error even after installing llama_cpp_agent :
(moa) Ubuntu@0136-ict-prxmx50056:~/Mycomind-Daemon-Ollama-Mixture-of-Memory-RAG-Agents$ python omoa.py Traceback (mo…
-
Hi thanks for all you great work at providing a wrapper with a web server. I got the wrapper working on my cpu but I have a ROCm system. I have llama.cpp fully working on my GPU so I have tried to com…
-
We really on OpenAI API calls in two key areas:
1) To generate the distributions we create a prompt and then send it to the API for a response. Ideally we could just swap out this API call for a ca…
-
I run this on GPU: 2 * A30 with CUDA driver 535.104.12.
The docker image is built using `make -C docker release_build CUDA_ARCHS="80-real"`
I use the latest code in branch main.
```
commit 89ba1…