Open sgwhat opened 6 months ago
@sgwhat https://github.com/ollama/ollama/blob/main/docs/api.md#load-a-model
This might be a good solution
@sgwhat https://github.com/ollama/ollama/blob/main/docs/api.md#load-a-model
This might be a good solution
I have tired, but still useless.
Validations
Problem
I've noticed that when I use ollama to chat, the model loading always occurs during the first round of conversation, which makes the first round much slower than subsequent ones.
I'm exploring ways to help ollama preload the model. Even though I tried
ollama run llama2:latest
before conversation, the model still loads at the start of the first conversation.Solution
No response