-
### Package:
https://aur.archlinux.org/pkgbase/llama.cpp-git
### Purpose:
llama.cpp is a port of Facebook's llama model in C/C++, and supports running many large language models (llms).
This pkg…
jlo62 updated
3 months ago
-
Hi!
I use the lib rocblas.for.gfx90c.workable that works perfectly on my Ryzen 7 5700g, it has worked without problems in Comfy and Stable Difussion, but when I try to use it in Text-Generation-Web…
-
Hi, i tried finetuning both llama 3.1-8b-instruct and llama 3-8b-instruct following the notebook you provided [here](https://colab.research.google.com/drive/1XamvWYinY6FOSX9GLvnqSjjsNflxdhNc?usp=shari…
-
Hi guys, I'm getting these errors when trying manually install this:
Windows 11 64-bit. Thanks for any help.
PowerShell 7.4.5
PS C:\Windows\System32> python -m pip install https://github.com/oob…
-
On adding llama_cpp-rs to my Cargo.toml, llama.cpp seems to be locked to an older version. I'm trying to use Phi-3 128k in a project and I'm unable to because the [PR that was merged into llama.cpp](h…
-
Got this while running from main branch in Podman AI Lab:
```
�llama_model_loader: loaded meta data with 25 key-value pairs and 291 tensors from /granite-7b-lab-Q4_K_M.gguf (version GGUF V3 (lates…
-
# Overview
This is a list of changes to the public interface of the `llama` library. Collaborators are encouraged to edit this post in order to reflect important changes to the API that end up merg…
-
how run GOT-OCR2.0 on cpu only?
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
**Qwen2**
warning: not compiled with GPU offload support, --n-gpu-layers option will be ignored
warning: see main README.md for information on enabling GPU BLAS support
Log start
main: build = 2…