Closed BearXP closed 1 hour ago
Yeah, a configuration file (yaml or json) that users can modify to specify if they want to use GPT4-o or Ollama and which model / server they want to use sounds like the best option for now.
If you can create a pull request for this and provide some simple instructions on how to test it, then I'd be grateful and very happy to merge it in.
Ended up implementing it myself 5ed1e9f6f00cb48a4288cf02b41db6670ba48213
Hello!
First I just wanted to say thank you for this, I saw your post on Reddit and this looks like such a fun project!
Looking through the code, it looks modular enough that it shouldn't be too hard to add support for a local model like Ollama.
pip install ollama
In
util.py
add the following:Then in
aisnip.py
addOllamaModelWrapper
.The hardest bit left would be getting the users input for where their ollama server is running (I've hard coded it to localhost), and which model they want to use (again I've just hard coded it to minicpm-v).
If you want I'm happy to setup an ollama_config.env or something and learn how to do a pull request?