-
Hey everyone!
Just opening an issue to track which models people would like to see supported with Serge.
- [x] Alpaca 7B, 13B & 30B
- [x] Alpaca 7B-Native
- [x] gpt4all
- [x] Vicuna 7B & 13B
-…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [X ] I am running the latest code. Development is very rapid so there are no tagged versions as o…
-
**Describe the bug**
I'm using LM Studio with MemGPT, and as I seeing and followed some tutorials.
**To Reproduce**
Steps to reproduce the behavior:
On the memgpt enviroment inside the folder …
-
I have been using Zep for a while and I have found it to be fantastic.
Setup:
Conversational Local LLM through Langchain
Local Embedding on Zep
OpenAI gpt 3.5 for Zep's LLM.
However, I have no…
-
How to install it?
What this mean?
CMD_FLAGS --model-menu --model Wizard-Vicuna-13B-Uncensored-GPTQ --model_type llama --wbits 4 --groupsize 128 --chat --no-stream api --listen-port 7861 --listen
…
-
I am on macOS. I run `ollama run llama2-uncensored:70b` and get the following:
```sh
pulling manifest
pulling 47f73cb430c8... 100% |███████████████████████████████████████████████████████████████…
-
### Describe the bug
Hi,
When loading a model (TheBloke/WizardLM-33B-V1.0-Uncensored-SuperHOT-8K-GPTQ) with ExLlama, the system only uses 1 GPU.
I got 3.
1x RTX 4000 8Gb
2x RTX A4000 16Gb
I …
-
I've noticed that after running a few models, sometimes the models don't behave normally. This is a session where that was occurring. I had first tried with bakllava but it wasn't being helpful eithe…
-
All I need is to run ollama3 on an Intel GPU (Arc™ A750) and I follow the steps as described in the IPEX-LLM documentation, but it runs on the CPU. Search engines can't find a solution to the problem.…
-
### Description
So after the 4-5 days of forum downtime, there seem to be a lot of changes to the post filter, which in turn blocks posting image links created by the `image-uploader` addon.
https…