Open r0kk opened 1 week ago
So the examples we used following Ollama OpenAI API specs https://github.com/ollama/ollama/blob/main/docs/openai.md#curl
Please use the Test connection feature to make sure the Ollama connection working properly for both LLM & embedding models.
It worked before, but ollama
changed endpoints from /embeddings
to /embed
, so OpenAI
client should not work anymore because it uses /embeddings
. At least this is my understanding.
The openAI client endpoint:
Same issue here with ollama -v ollama version is 0.4.1 and kotaemon full docker image
+1
Is this just an issue with Ollama v0.4?
ollama -v
ollama version is 0.3.9
Call from within Kotaemon app docker runtime:
root@justin-two-towers:/app# curl localhost:11434/api/embeddings -d '{ "model": "llama3.1:8b", "input": "Why is the sky blue?" }'
{"embedding":[]}
Seems fine ...
Description
Hi, I think you are calling the wrong endpoint for local embedding for ollama, if I use settings from your instructions here
From official ollama api documentation here it should be called:
http://localhost:11434/api/embedd
endpoint, but from koteamon it is calledhttp://localhost:11434/api/embeddings
following works: curl http://localhost:11434/api/embedd -d '{ "model": "",
"input": "Why is the sky blue?"
}'
following does not: curl http://localhost:11434/api/embeddings-d '{ "model": "",
"input": "Why is the sky blue?"
}'
There is also a problem that on UI we don't get any notification about the issues and one has to look into the logs. It would be great if it could be a little bit more explicit.
Reproduction steps
Screenshots
Logs
No response
Browsers
No response
OS
No response
Additional information
No response