Closed wwqq closed 10 months ago
Could you provide the complete log like this? #13
It looks like the models are not loaded successfully. There are two possible reasons:
git lfs pull
is recommended while downloading the model from huggingface. After downloading the model, you can try the inference with python script first if possible.tensor mm.model.mlp.1.weight
lies in mmproj-model-f16.gguf
.
To convert the model to GGUF, you must have deployed llama.cpp
on you PC. You can test the model on you PC first to check if the model is converted successfully then run it on android later. Please make sure all of paths of models are right and the args of command line are in right format.Later we will upload the GGUF model of MobileVLM somewhere and update a simpler instruction on android deployment.
git lfs pull
it works. thx
I follow the deployment instructions https://github.com/XiaotaoChen/llama.cpp/blob/MobileVLM/examples/llava/MobileVLM-README.md step-by-step. But when I ran the command
The output file is:
How to solve this problem?Thanks!