-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
# Expected Behavior
I was trying to just install and use llama-cpp-python on a freshly set up anaconda environment.
I updated my anaconda navigator to the 2023.09 release version.
I created a new…
-
### System Info / 系統信息
cuda 11.8
llama-cpp-python 0.2.55
python 3.10
windows 10
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [X] pip install …
-
The version of llama-cpp-python this project uses is quite old. Therefore I get a lot of errors regarding versions of GGML models. It also doesn't support GGUF models.
I would suggest to up the ver…
-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
Hi,
I need clarification on the `is_chat_model` parameter in `LLMMetadata`. Specifica…
-
Having 32 GB of GPU and 64GB of ram intel 17 13th gen processor its taking 2-4 min to response and not using GPU using llama-cpp-python==0.1.83 --no-cache-dir
![image](https://github.com/PromtEngine…
-
**Installation command (conda environment):**
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python
**produces the following output:**
Collecting llama-cpp-python
Usi…
-
I am trying to use this incredible new library you are working on. And following your example in the arxiv notebook I received this error:
![image](https://github.com/user-attachments/assets/cb0c92ba…
-
I'm attempting to install llama-cpp-python under the tensorflow-gpu docker image (nightly build) . When I attempt to do so, I get the following error messages.
````
root@a1f1e127514b:/tf# CMAKE_A…
-
## Describe the bug
When i add a new offline ,odel and start the chat with any input., i can see that khoj starts to download the model in the console. but after it finished downloading the model, kh…