Closed starsoccer closed 1 month ago
Thank you so much! I have experienced this myself but thought it had to be something else. Will be fixed in v1.2.1.
Awesome I can confirm this is working no. For anyone else who sets this though be aware that atleast with ollama sentences can be cut off so it is worth updating the prompt to also include the maximum response you want. It would be nice iif there was a way to do this automatically as well.
It seems ollama does not support a max_tokens parameter and instead I think expects a
num_predict
parameter instead based on the docs here, https://github.com/ollama/ollama/blob/main/docs/modelfile.md#valid-parameters-and-valuesCurrently when using LLM Vision I get the below warnings in Ollama because max_tokens is set: