We've been testing running assistants with ollama. This is of interest for some folks using agency-swarm with assistants but may be interesting to others as well (assistants-api with local models).
Currently this endpoint seems to stream even though the stream argument is not set to true:
response = client.chat.completions.create(
model="ollama/phi3",
messages=[{"role": "user", "content": "respond in 20 words who are you"}]
)
We've been testing running assistants with ollama. This is of interest for some folks using agency-swarm with assistants but may be interesting to others as well (assistants-api with local models).
Currently this endpoint seems to stream even though the stream argument is not set to true:
sample response:
Related discord chat here: https://discord.com/channels/1245465949679915008/1245478014737977344/1255181994317451346