Closed Kreijstal closed 7 months ago
ollama appears to provide a custom api and only /v1/chat/completions. Polymind only supports llama.cpp's official server, tabbyapi and openai compatible /v1/completions endpoints, so unless you can get ollama to serve a standard completions endpoint, no.
https://github.com/ollama/ollama/pull/5209 So it's possible right...!!!
I tried to change the port to something llama serve listens to and I get this