-
**Describe the bug**
**To Reproduce**
Steps to reproduce the behavior:
excute this command:
CMAKE_ARGS="-DLLAMA_CUDA=on -DLLAMA_NATIVE=off" pip install 'instructlab[cuda]'
and compile err…
-
We want to observe interactions of llama-cpp, try to get inspiration from https://github.com/cfahlgren1/observers/blob/main/src/observers/observers/models/openai.py
```python
from llama_cpp import…
-
OS: 22.04.1-Ubuntu
Python: Python 3.12.2
Build fails for llama-cpp-python
```
$ pip install -r requirements.txt
...
Building wheels for collected packages: llama-cpp-python
Building wheel…
-
Was told to move this https://github.com/comfyanonymous/ComfyUI/issues/5510 story to ComfyUI-N-Nodes repo.
### Expected Behavior
I'm not sure if Ollama models are required in anyway but I do see…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
### Describe the issue as clearly as possible:
Consider the following code:
```
from outlines import models
# No error
models.llamacpp(
repo_id="M4-ai/TinyMistral-248M-v2-Instruct-GGUF"…
-
After Ver0.3.0, "eval time" and "prompt eval time" of llama_print_timings are displayed as 0.00ms.
I thought it was a problem with llama.cpp, but it was displayed correctly.
Here is a code and r…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ Yes] I am running the latest code. Development is very rapid so there are no tagged versions as…
-
There are multiple issues with the CUDA wheels:
1. The cu125 repository returns 404:
```bash
$ curl -I https://abetlen.github.io/llama-cpp-python/whl/cu125/
HTTP/2 404
```
2…
-
**Chapter 6**
I am running the below in Colab connected to T4.
%%capture
!pip install langchain>=0.1.17 openai>=1.13.3 langchain_openai>=0.1.6 transformers>=4.40.1 datasets>=2.18.0 accelerate>=0.…