Open not-lain opened 3 weeks ago
I found something interesting
If you go to https://api.github.com/repos/Mozilla-Ocho/llamafile/releases/latest and check the ["assets"][0]
you will find the ["name"]
as well as the ["browser_download_url"]
there, which can be used to automatically update the snippets
llamafile is a local app (similar to llama.cpp) to run llms in a distributed way from a single file
library can be used on both
.gguf
and.llamafile
filesrepo : https://github.com/Mozilla-Ocho/llamafile
snippets
linux and mac
windows (download and rename it using .exe)
gguf
notes
.\llamafile-0.8.13 -m foo.llamafile
to get around the limit (similar to GGUF snippet)./llava-v1.5-7b-q4.llamafile
it'll launch an HTTP server, open a tab in your desktop's browser, and you can chat with the model, upload an image file, ask it to analyze what it sees, etc.