Closed netnem closed 7 months ago
Thanks for your thorough report, hope to address this soon as I've seen this error in another context, and I can reproduce it now.
I'm fixing this right now
FYI I seem to have issues with ollama independent of LibreChat when I don't include --drop_params
, related issue: https://github.com/BerriAI/litellm/issues/992#issuecomment-1839932916
Confirmed the latest merge works great.
Thanks @danny-avila
Contact Details
No response
What happened?
I have Ollama/openchat running behind the OpenAPI compatible frontend of LiteLLM.
The chat completion never "finishes" when the bot is responding, appears to be searching for a chatcompletion message
Steps to Reproduce
install librechat via docker, in .env file set OPENAI_REVERSE_PROXY=http://192.168.2.142:8000 (i installed litellm on the docker host) install ollama via install script: curl https://ollama.ai/install.sh | sh pip install litellm ollama pull openchat pip install async_generator litellm --model ollama/openchat --api_base http://localhost:11434 --drop_params <-- required for presence penalty not being supported in litellm
The setup works for calling a localLLM, but the cursor gets "stuck" and does not return to the user to add additional chats.
What browsers are you seeing the problem on?
Firefox, Microsoft Edge
Relevant log output
Screenshots
Code of Conduct