I ran llamafile without any flags, it loaded the web UI with CPU inference. I added the GPU flags (--gpu nvidia -ngl 999) and got a segfault. I installed CUDA (then verified with a "Hello, world!"), re-ran the same command and got the same segfault.
I removed ~/.llamafile and reran with the GPU flags and everything worked.
PS: I looked through the issue tracker and couldn't find a duplicate, but maybe this problem is mentioned in a comment thread.
Contact Details
No response
What happened?
I ran llamafile without any flags, it loaded the web UI with CPU inference. I added the GPU flags (
--gpu nvidia -ngl 999
) and got a segfault. I installed CUDA (then verified with a "Hello, world!"), re-ran the same command and got the same segfault.I removed
~/.llamafile
and reran with the GPU flags and everything worked.PS: I looked through the issue tracker and couldn't find a duplicate, but maybe this problem is mentioned in a comment thread.
Version
$ ./llava-v1.5-7b-q4.llamafile --version llamafile v0.8.11
What operating system are you seeing the problem on?
Linux
Relevant log output