Open jtpio opened 2 weeks ago
I think it's a mix of all of these.
Maybe we could make it more reliable with a better use of the inline completer and the throttler. I'm thinking about:
the Mistral API seems to take around 2s to respond on my hand (and sometime the first requests never ends).
2s sounds like a lot. Wondering if we can already see some improvements with other providers like in https://github.com/jupyterlite/jupyterlab-codestral/pull/17 for example. If that's the case maybe there isn't much we can do about the Mistral provider.
the Mistral API seems to take around 2s to respond on my hand (and sometime the first requests never ends).
2s sounds like a lot. Wondering if we can already see some improvements with other providers like in #17 for example. If that's the case maybe there isn't much we can do about the Mistral provider.
As far as I know, Groq does not provide a completion model out of the box (not with langchain at least).
OK, we could try with OpenAI then.
When trying https://github.com/jupyterlite/jupyterlab-codestral/pull/15 locally with MistralAI, the inline completer seems to be "off" most of the time, making it difficult to use:
I believe this was already noticed in https://github.com/jupyterlite/jupyterlab-codestral/pull/8 and may be the same issue.
It's not clear yet whether it's because the Mistral API is slow to respond, because of its 1 req/s limit or how the inline completer is currently set up.