-
Servers stops responding after one API call, and chat starts streaming ``.... or on gemma replies only with GGGGGGGGGs
-
I am seeing below error when I am trying to chat. I tried deleting the database and running "python manage.py runserver" command but still seeing the same issue.
Any suggestions to resolve the issu…
-
When I run glm-4-9b-chat-Q5_K_M.gguf on the Cuda 12 machine, the API server can be started successfully. However, when I send a question, the API server will crash.
The command I used to start the …
-
it should send /nonserver_chat information for metrics to end up on axiom
then we can see free users vs subscribed users metrics
-
### Issues Policy acknowledgement
- [X] I have read and agree to submit bug reports in accordance with the [issues policy](https://www.github.com/mlflow/mlflow/blob/master/ISSUE_POLICY.md)
### Where…
-
I need to look more closely at error handling
-
I have tried many conversation with memGPT agent using https://inference.memgpt.ai .
It sounds good, at the beginning
But at some point of the conversation, after about 10 or 20 exchange between u…
-
After opening chat app its show this error message. Autocompletion works normaly.
``` json
{
"url": "http://127.0.0.1:59791/?api_key=MY API KEY",
"connection": {
"status": "CONNECTING",
…
-
I am trying to explore the backend server. After resolving dependencies issues, I tried to start the server but system doesn’t shows any running backend server neither logs helps out to identify the i…
-
**Please describe the feature you want**
Related: https://github.com/TabbyML/tabby/issues/2652
This allows a local deployment could use fewer vram / computing in local setup
**Additional co…