Open juhp opened 1 week ago
Or is this because I don't have ollama (etc) installed locally say?
Looking at roughly model.py line 101. It's because you don't have llama.cpp installed on your system, looking at the code, it's trying to exec llama-server or llama-cli and you don't have it on your system:
Users who don't use containers are gonna run into this kinda thing all the time. These AI frameworks are complex, --nocontainer is for advanced Linux users and macOS users.
Note if you want to use with a GPU, you will likely have to build llama.cpp in a very specific way.
Something that could be a nice contribution is to make this error message clearer, so if someone else tries this in the future they have a more informative message (they can see it's specifically llama-server/llama-cli that it is trying to execute and failing on).
@juhp we don't use Ollama directly, we have our own implementation of an Ollama client written from scratch.
When I try run with
--nocontainer
I get a backtrace:This is using an install with my modified spec file which is in my test copr
(I first hit this trying to run ramalama inside a toolbox container, but also get it in a vm with
--nocontainer
.)