Closed rhelmot closed 9 months ago
Hello @rhelmot, thank you for your contribution. What OpenAI-compatible frontend are we talking about here?
I took a quick glance, it looks good. I'll take a deeper look when I have a bit more time.
I was using python -m llama_cpp.server, but I believe there are more as well. litellm comes to mind.
Closing in favor of #40
With this change and some minor tweaks to llm.nvim, I was able to get code completions from a locally running copy of CodeLlama-13b through an OpenAI-compatible frontend.