Closed francesco086 closed 4 months ago
👀 @francesco086
Thank you for raising an issue. We will investigate into the matter and get back to you as soon as possible. Please make sure you have given us as much context as possible.\ 非常感谢您提交 issue。我们会尽快调查此事,并尽快回复您。 请确保您已经提供了尽可能多的背景信息。
@francesco086 I think it might be the issue of parameter. Try to add a high value of frequency_penalty
.
💻 Operating System
Other
📦 Environment
Docker
🌐 Browser
Other
🐛 Bug Description
I serve the mistral-7b model using ollama, and set a litellm proxy in front of it. I am, for example, able to run the command:
and get the expected response
I setup lobechat to use several OpenAI models via litellm (gpt 3.5, 4, and dalle3), and everything works fine. However, with ollama/mistral-7b I get the following behaviour (I pressed the "Stop" button after a while because it was too slow):
🚦 Expected Behavior
No response
📷 Recurrence Steps
No response
📝 Additional Information
Services are running on Kubernetes, setup via ArgoCD.