Open victorserbu2709 opened 1 day ago
From https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#api-reference
Note: parallel_tool_calls and user parameters are ignored.
Sorry I'm a bit confused - do you expect vLLM to return multiple tool calls? Unfortunately letta doesn't support parallel tool calling at the moment.
Is your feature request related to a problem? Please describe. Hello. I tried to use letta with vllm serving qwen2.5 72B model. It returned 2 tools and letta doesn't support this
Describe the solution you'd like To work
Describe alternatives you've considered A clear and concise description of any alternative solutions or features you've considered.
Additional context Add any other context or screenshots about the feature request here.