Mozilla-Ocho / llamafile

Distribute and run LLMs with a single file.
https://llamafile.ai
Other
19.39k stars 982 forks source link

run-detectors: unable to find an interpreter for ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile #369

Closed superkuh closed 5 months ago

superkuh commented 5 months ago

Edit: nevermind, this is covered in https://github.com/Mozilla-Ocho/llamafile?tab=readme-ov-file#gotchas

Hi. I downloaded Meta-Llama-3-8B-Instruct.Q6_K.llamafile and chmod +x it and attempted to run it I but I get an error.

superkuh@janus:~/app_installs/llama.cpp/models$ ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile -ngl 9999 run-detectors: unable to find an interpreter for ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile

superkuh@janus:~/app_installs/llama.cpp/models$ ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile --gpu disable run-detectors: unable to find an interpreter for ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile

superkuh@janus:~/app_installs/llama.cpp/models$ ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile run-detectors: unable to find an interpreter for ./Meta-Llama-3-8B-Instruct.Q6_K.llamafile

My computer is a bog standard Ryzen 5 3600 running Debian 11. Since this fails in CPU mode too it's probably not related to my GPU, an AMD RX 580 8GB. Anyone know if there's a fix for this or even what's going on?

Using plain non-APE non-llamafile llama.cpp ./main and ./server work fine as an aside.