Open rpodcast opened 3 weeks ago
Building on this, it would be great if it was easy to switch to other providers (specifically azure endpoints). thanks
Turns out it my use case is really straight forward using the litellm docs https://docs.litellm.ai/docs/providers/azure :)
And I think this is how it would work for local models - https://docs.litellm.ai/docs/providers/ollama#using-ollama-apichat
@rpodcast in theory that link from Iain should do it. In practice, a couple of problems have emerged for me in the past when trying different providers.
@jcheng5 very cool stuff here! I've had a similar experience with llama3.1:8b and similar smaller models doing less well even just on sql query generation, though some of the smaller SQL-focused models like https://ollama.com/library/duckdb-nsql seem to do pretty well.
Maybe a separate issue, but in addition to local model support I'd be curious to hear if you've tried any of the other providers of interfaces? litellm
is new to me but looks very nice. I've been playing around with langchain and more recently with pandas-ai, both of which I think have more specific tooling for handling text->SQL->code execution pattern you are also using here? Would be super curious to compare notes!
I may have misunderstood when this was first presented to me, but are there plans to allow developers to bring their own local (self-hosted) LLM instead of leveraging openAI? That will be important for what I have in mind to try this out.