-
Hi everyone !
I have spent a lot of time trying to install llama-cpp-python with GPU support.
I need your help. I'll keep monitoring the thread and if I need to try other options and provide in…
-
### Describe the bug
I have recently updated the requirements of text-generation-webui and deployed the GGUF format of Mistral model using loader llama.cpp. I want to hit the webUI using langchain an…
-
### Describe the bug
I believe this is a new architecture based on [gemma-2](https://huggingface.co/google/gemma-2-9b-it) family model.
Error message:
```
llama_model_load: error loading model: er…
-
### Environment
🪟 Windows
### System
Win10
### Version
1.12.0
### Desktop Information
_No response_
### Describe the problem
![image](https://github.com/SillyTavern/SillyTavern/assets/15965…
-
I was wondering if Intel ARC Gpu's work with this. Could not read anything about this here.
-
### Before submitting your bug report
- [ ] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [ ] I'm not able to find an [open issue]…
-
Hi,
Thank you for creating this great web UI.
It's been great so far to run models locally on my machine.
I'm wondering if I could use this as an interface for API requests instead of running m…
-
### Describe the bug
`curl -v http://127.0.0.1:8000/v1/chat/completions …
-
Hi!
First of all thanks for the nice repo!
I tried already many different proposed solutions from here:
https://github.com/oobabooga/text-generation-webui/issues/5885
but I always get eith…
-
https://github.com/THUDM/ChatGLM-6B
https://huggingface.co/THUDM/chatglm-6b-int4
It would be best if you could provide some guidance. Thank you.