-
```
~►llama-assistant (base) 113.203s 15:13
Traceback (most recent call last):
…
-
Building wheel for llama-cpp-python (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building wheel for llama-cpp-python (pyproject.toml) did not run successfully.
│ exit co…
-
Some of features like ggml_graph_plan function was removed from ggml library when combining llama.cpp branch, it seems not using the previous branch from ggml with whisper.cpp is fully supported.
I…
-
When loading the model I get the following error message:
llm_load_tensors: ggml ctx size = 0.16 MB
llm_load_tensors: using CUDA for GPU acceleration
llm_load_tensors: mem required = 5734.11 MB
…
-
# Description
When attempting to set up llama cpp python for GPU support using CUDA toolkit, following the documented steps, the initialization of the llama-cpp model fails with an access violation…
-
### What happened?
When trying to run [FatLlama-1.7T-Instruct](https://huggingface.co/RichardErkhov/FATLLAMA-1.7T-Instruct) llama.cpp crashes while loading the model with the error: `n > N_MAX: 525…
-
### System Info / 系統信息
python版本:3.10.15
xinference版本:0.16.2
llama_cpp_python:0.3.1
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [X] pip install / 通过 pi…
-
I attempted to run a low-level API in version 0.2.11, but both installing from pypi and compiling from source failed.
python: 3.10.12
llama_cpp_python: 0.2.11
```bash
{llama-cpp-python/examples/lo…
islwx updated
2 months ago
-
# Expected Behavior
destructor llama normally
# Current Behavior
# source code:
import bs4
from langchain import hub
from langchain_community.document_loaders import WebBaseLoader
from langch…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…