Closed norteo closed 1 month ago
If you don't have an GPU, you probably don't need to parse -ngl 9999
.
This should only be an issue in older versions of llamafile. What version of llamafile is this llamafile associated with? To find out, you can run
./mistral-7b-instruct-v0.2.Q5_K_M.llamafile --version
Thank you for the reply.
user@fe9e8ccdc306:~$ ./mistral-7b-instruct-v0.2.Q5_K_M.llamafile --version
llamafile v0.8.0
It seems I was not using the latest version. I redownloaded the file and I rerun the curl command and it seems to work fine. The version I am using now is 0.8.5 . I just downloaded it from https://huggingface.co/Mozilla/Mistral-7B-Instruct-v0.2-llamafile/resolve/main/mistral-7b-instruct-v0.2.Q5_K_M.llamafile?download=true The hugging face repository does not seem to have the latest version and it seems it is not possible to know what llamafile version you are downloading. Maybe something should be done about that?
The hugging face repository does not seem to have the latest version and it seems it is not possible to know what llamafile version you are downloading. Maybe something should be done about that?
Would you be willing to post this as a separate issue?
I run llamafile with the mistral model as:
./mistral-7b-instruct-v0.2.Q5_K_M.llamafile -ngl 9999 --port 8080 --host 0.0.0.0 --embedding --threads 16
I don't have a GPU.
If I run
llamafile "crashes" with the message: