Closed RayRama closed 4 months ago
I believe the model in question could not generate correct output structure - that's why validation failed. My suggestion would be to try a different model or a different structure mode ('auto' | 'json' | 'tool' | 'grammar'
)
As someone new to the repo it ends up being a bit confusing. On one hand the examples or documentation doesn't immediately suggest popular models like ollama ->llama3 would be running into problems.
Of course its understandable to not keep track of every single model that is out there, but for self hosted options perhaps recommendations could be given of what is tested to at least be a starting point.
for example I tried all 4 structure modes and none of them worked with llama3:8b....if that is a dead end and this model doesn't work with the tool how do we go about figuring out which does work apart from trial and error?
Try the new llama3.1, it has built-in tool use and in my experience generates correct json most of the time
When I try to use llama3 locally, I got some error like this
I used barely example from the readme file and changed the LLM provider to llama3