Open bitnom opened 4 months ago
I understand the request and I think OAI did a great job with their function calling models. I am uncertain if support for this in LMQL makes sense, since it is a very vendor-specific API, that will be hard generalize in a model-agnostic way.
I am open to design proposal however. It would be great to somehow abstract their implementation away, and to provide a common interface, that also works e.g. for Gorilla models or other forms of more open function calling.
I understand the request and I think OAI did a great job with their function calling models. I am uncertain if support for this in LMQL makes sense, since it is a very vendor-specific API, that will be hard generalize in a model-agnostic way.
I am open to design proposal however. It would be great to somehow abstract their implementation away, and to provide a common interface, that also works e.g. for Gorilla models or other forms of more open function calling.
I did a bit more research on this issue. There are several major implementations of function/tool calling:
There are of course others. Every agentic framework has some version of this. All of them I have looked at so far have standardized around the OpenAI parallel function calling API, and use the openai
Python package for integration, both for calls to openai and function calling via local LLMs.
This standardization is not confined to agentic frameworks. The fact that Gorilla and litellm are depending on the openai spec for all function calling are pretty good signals.
I would have already submitted a PR but the lmql oai client is more complex than most other packages' openai integrations. The reason for that is obvious. Regardless, I have yet to have time to sit down and take it all in.
Note: As I vaguely recall from my first attempt, logit_bias
and/or logprobs
parameters did not have any effect on the output of a function/tool definition and/or call. I could be misremembering, so citation and/or example needed because that doesn't sound right. Streaming/chunks is supported.
Another thought: check the prompt templates of gorilla and litellm's function-calling for local models.
Investigate the feasibility of adding support for the OpenAI function-calling API, and (If feasible) integrate it.
Rationale
source: https://gorilla.cs.berkeley.edu/blogs/4_open_functions.html
References
https://platform.openai.com/docs/guides/function-calling
Example request/response:
Response:
Notes