-
I have been roleplaying with CommandR+, an 104b model. So far, I am using 40,000 out of 65,000 context with KoboldCPP. Considering that this model has been lucid so far, I am expecting to eventually…
-
At this point, given the >4GB allocation bug was fixed, if possible, I'd like to see the SYCL backend integrated into koboldcpp. While Intel GPUs are supported by Vulkan, SYCL is much faster at prompt…
-
otherarch/ggml_v3-cuda.cu(609): warning #177-D: function "warp_reduce_sum(half2)" was declared but never referenced
otherarch/ggml_v3-cuda.cu(630): warning #177-D: function "warp_reduce_max(half2)"…
-
```
def RunServerMultiThreaded(addr, port, embedded_kailite = None, embedded_kcpp_docs = None):
global exitcounter, sslvalid
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
so…
Kas1o updated
2 months ago
-
Kind of a noonish I really dig what your doing but open ai going to be adding quick when releasing to madness I want to use open source is there ways to do this only with the ollama
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ x] I am running the latest code. Development is very rapid so there are no tagged versions as o…
-
Hello!
thank you for this.
is there any chance getting a gui without needing to install python like koboldcpp in the future for its portability?
also, can we use this with other models like w…
-
### Feature request
Have ability to
1. edit last request, to get a better response quality
2. edit previous request/response to tune response quality
3. select and delete previous request/respon…
-
With the newest Version of the Ooba Booga Text Generation WebUI they replaced the old KoboldAI API with the new OpenAI compatible API. My Tavern Ai wasn't able to connect to this new API either with t…
-
Is it possible to connect the plugin via the oobabooga-webui or koboldcpp API to a locally running model (Refact-1.6b, starcoder, etc.)?
If possible, how? Or is it possible to work with local models …