-
I've looked through the issues, I do see there are some complaints about this, a lot of different ways to fix it working differently for different users.
Is there a unified, accepted and tested way…
-
```
~►llama-assistant (base) 113.203s 15:13
Traceback (most recent call last):
…
-
Building wheel for llama-cpp-python (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building wheel for llama-cpp-python (pyproject.toml) did not run successfully.
│ exit co…
-
When loading the model I get the following error message:
llm_load_tensors: ggml ctx size = 0.16 MB
llm_load_tensors: using CUDA for GPU acceleration
llm_load_tensors: mem required = 5734.11 MB
…
-
**Describe the bug**
Inserting a barrier inside `matmul/device/kernels/dataflow/reader_bmm_tile_layout_in0_sender_dram_sharded.cpp` after a `noc_async_write_multicast_loopback_src` causes DRAM Sharded…
-
I attempted to run a low-level API in version 0.2.11, but both installing from pypi and compiling from source failed.
python: 3.10.12
llama_cpp_python: 0.2.11
```bash
{llama-cpp-python/examples/lo…
islwx updated
2 months ago
-
### What happened?
When trying to run [FatLlama-1.7T-Instruct](https://huggingface.co/RichardErkhov/FATLLAMA-1.7T-Instruct) llama.cpp crashes while loading the model with the error: `n > N_MAX: 525…
-
### System Info / 系統信息
python版本:3.10.15
xinference版本:0.16.2
llama_cpp_python:0.3.1
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [X] pip install / 通过 pi…
-
# Description
When attempting to set up llama cpp python for GPU support using CUDA toolkit, following the documented steps, the initialization of the llama-cpp model fails with an access violation…
-
My ComfyUI is not the portable. I installed the Searge LLM with the ComfyUI manager, then I installed it manually, in both cases; I got the Traceback below error. I tried to Install the following comm…