-
### What happened?
Hi, im trying to use Google [Madlad400 in GGUF version,](https://huggingface.co/NikolayKozloff/madlad400-10b-mt-Q8_0-GGUF) but I'm unable to work it with `llama-server` but it work…
-
```
`Building wheels for collected packages: llama-cpp-python
Building wheel for llama-cpp-python (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building wheel for llama-…
-
### System Info / 系統信息
cuda 11.8
llama-cpp-python 0.2.55
python 3.10
windows 10
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [X] pip install …
-
Hello,
I installed privateGPT using:
```
salloc --job-name "InteractiveJob" --cpus-per-task 4 --mem-per-cpu 50gb --time 01:00:10 -p batch_gpu --gres=gpu:1
module load Mamba git CUDA && conda…
-
### Feature Description
Add support for Llama 3.1 8B, 70B, and if possible 405B models.
### Additional Context (optional)
_No response_
### Checklist:
- [X] I have checked for existing issues tha…
-
The Vicuna models got included in llama.cpp: https://github.com/ggerganov/llama.cpp/discussions/643#discussioncomment-5533894
With both quantized and unquantized weights being released: https://huggi…
-
# Expected Behavior
I was trying to just install and use llama-cpp-python on a freshly set up anaconda environment.
I updated my anaconda navigator to the 2023.09 release version.
I created a new…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
### Problem Description
Hello,
Im trying to use llamacpp with rocm6 and hipBlas under pcie x1 gen1 without success.
So im using at the moment Vulkan
### Operating System
"22.04.4 LTS (Jammy Jel…
-
The return value of the solver part of `evanix` should contain enough information to generate a report saying (0) how many derivations in total will be built (in future versions also how many compute …