⏩ Continue is the leading open-source AI code assistant. You can connect any models and any context to build custom autocomplete and chat experiences inside VS Code and JetBrains
Start ollama serve with OLLAMA_KEEP_ALIVE=-1 environment variable.
Chat inside Conitnue Extension.
Ollama release model memory after a while instead of keeping it in memory.
Before submitting your bug report
Relevant environment info
Description
OLLAMA_KEEP_ALIVE not working with continue extension. See https://github.com/ollama/ollama/issues/5272#issuecomment-2489940273 and following comment.
To reproduce
Start ollama serve with OLLAMA_KEEP_ALIVE=-1 environment variable. Chat inside Conitnue Extension. Ollama release model memory after a while instead of keeping it in memory.
Log output
No response