-
Am I doing this correctly?
```
D:\textgen\kobold>.\koboldcpp.exe --useclblast 0 0 --smartcontext
Welcome to KoboldCpp - Version 1.10
Attempting to use CLBlast library for faster prompt ingestio…
-
How i build:
1. I use [w64devkit](https://github.com/skeeto/w64devkit/releases)
2. I download [CLBlast](https://github.com/CNugteren/CLBlast) and [OpenCL-SDK](https://github.com/KhronosGroup/Open…
-
# Expected Behavior
using API-key and be able to turn on share with horde
# Current Behavior
option not there
would love to be able to use as worker so koboldcpp becomes multi-user
…
-
So I'm trying to do some SFT, but whenever I train it ends up inserting a space or some sort of blank character right before the response. Happens with the templates I tried like Vicuna and Alpaca. Li…
xzuyn updated
8 months ago
-
I run `koboldcpp.exe`, wait till it asks to import model and after selecting model it just crashes with these logs:
![logs](https://i.imgur.com/Cz64Kxn.png)
I am running Windows 8.1 with 8 GB of…
-
Every time I advance a little in my discussions, I crash with the following error:
`Processing Prompt [BLAS] (1024 / 1301 tokens)ggml_new_tensor_impl: not enough space in the scratch memory`
My …
-
So i've been getting back into IT after years away and dabbling with the AI models. I successful ran LLaMa.cpp in win64devkit the other day, although very slowly. I am trying to now run the same model…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [X] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
I've used lite.koboldai.net for the past 1~2 weeks, as well as running a worker off-and-on via KoboldAI (version: 0cc4m/koboldai : latestgptq)
(Hardware: GTX 1070 (8GB))
And have noticed some stra…
-
I've been trying to use koboldcpp with a 200 token limit, and I've noticed that every model defaults back to generating conversations with itself to fill the set limit, even when I have multiline resp…