Open gfxenjoyer opened 1 month ago
If you dont mind, i tried doing the ollama inference, but cant get to execute the stack build. can you suggest me what steps you followed ?
I'm also interested to hear how we can run any ollama models?
cant get to execute the stack build. can you suggest me what steps you followed ?
Are you able to share the errors you are seeing? You can follow the getting started guide: https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.md
I'm also interested to hear how we can run any ollama models?
Curious to understand your use case. Can you share some details?
Why we cannot run orher models like llama3.2 when using ollama? {"message": "400: Invalid value: Could not find provider for Llama3.1-8B-Instruct"}}
. Only Llama3.1-8B-Instruct works when using ollama as inference.
I managed to run other models like
gemma2
andphi3.5
by changing the linesin
./llama_stack/providers/adapters/inference/ollama.py
.Can the Ollama provider give us an escape hatch to easily run Ollama models?
Thanks