-
It would be great if support for the SDXS-512 model could be added: https://github.com/IDKiro/sdxs
Especially for quickly generating images on the CPU, this is a major new development, so it would …
-
The app fails while loading models:
```
root@koboldcpp-cb947d9b7-jqrfp:/koboldcpp# python koboldcpp.py --model /app/models/mythalion-13b.Q8_0.gguf --threads 8 --noavx2 --debugmode
***
Welcome to K…
-
[kobold_debug.json](https://github.com/henk717/KoboldAI/files/15272513/kobold_debug.json)
For some reason token streaming just does not work. It's enabled and the actual terminal output from the se…
-
With the newest Version of the Ooba Booga Text Generation WebUI they replaced the old KoboldAI API with the new OpenAI compatible API. My Tavern Ai wasn't able to connect to this new API either with t…
-
I was helping someone else get this working over the last couple of days and it struck me that it would be awesome if there were precompiled binaries with / without cuda for windows / mac / linux the …
-
```
Processing Prompt (1 / 1 tokens)
Generating (1 / 242 tokens)
(EOS token triggered! ID:2)
CtxLimit: 3768/4096, Process:0.44s (442.0ms/T = 2.26T/s), Generate:0.00s (3.0ms/T = 333.33T/s), Total:0…
-
How to setup gpu on termux setup? I hope it works I'm about to test sdxl open dalle v1.1 but ik I'll have to use sdquant command I got enough ram but just to speed up the process.
BTW is it possibl…
-
Kind of a noonish I really dig what your doing but open ai going to be adding quick when releasing to madness I want to use open source is there ways to do this only with the ollama
-
Hello.
Usually, if --quiet is not set, we usually get this during inference:
generating: 12/512 tokens
but this also outputs the prompts and the response. On the other hand, if --quiet is s…
-
My Computer:
Operating System: Windows 11 Professional 22H2 (Chinese Traditional)
CPU: AMD Ryzen 5 5600 6-Core Processor 3.50 GHz
GPU: GIGABYTE AMD Radeon RX 6650 XT (Driver Version 23.10.2)
RAM: …