-
Describe the bug:
Error for memgpt talking to local LLM.
Please describe your setup:
What is the output of memgpt version? (eg "0.2.4")
0.3.7
How did you install memgpt?
pip install pymemgpt…
-
### Describe the bug
When installing on windows with option NVIDIA (A) and choosing CUDA 11.8 Support (Y), the installation is failing because of:
ERROR: HTTP error 404 while getting https://githu…
-
I am trying to launch llama-2 from the [oobabooga_macos repo](https://github.com/oobabooga/text-generation-webui) but am encountering errors on my MacOS as stated below:
ERROR: Failed building wheel…
-
### Describe the bug
It just fails and uses the CPU instead, not sure what the issue is. I use oobabooga separately and can load models with gpu offloading via llama.cpp, so I don't know. Any suggest…
-
### Describe the bug
idk its bug or im dumb or else if im trying 7B model AWQ on Nvidia Quadro M4000 and i5-13600K it starts im loading model everything works but responses from chat is blank it says…
-
### Have you searched for similar requests?
None
### Is your feature request related to a problem? If so, please describe.
Everytime i make a new message in the chat the time to generate gets longe…
-
It seems that an update is needed.
https://github.com/oobabooga/text-generation-webui/pull/1538/files
-
### Describe the bug
start_linux.sh fails
The following packages were not installed with ./start_linux.sh and had to be installed manually into the ${HOME}/src/text-generation-webui/installer_file…
-
So I have a GPTQ llama model I downloaded (from TheBloke), and it's already 4 bit quantized. I have to pass in False for the load_in_4bit parameter of:
```
model, tokenizer = FastLlamaModel.from_pr…
-
**Description**
About 10 days ago, KoboldCpp added a feature called Context Shifting which is supposed to greatly reduce reprocessing. Here is their official description of the feature:
> NEW FE…