Closed JennieGao-njust closed 9 months ago
We exploit https://github.com/ggerganov/llama.cpp as the inference framework for mobile phone and Jetson AGX Orin platform.
The specific deployment plan is still being organized and will be released in one or two weeks. We will release all of the necessary technical details, coming soon with our MobileVLM-v2.
We have updated README.md with the deployment instruction on Android devices with customized llama.cpp
. Have a try!
Hi, we are closing this issue due to the inactivity. Hope your question has been resolved. If you have any further concerns, please feel free to re-open it or open a new issue. Thanks!
mobileVLM 实现移动端部署的路径是什么呢,我不太了解模型的mobile部署方式,是用什么框架推理呢,mnn或者fastllm可以用吗