TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and performance optimization for mobile devices, and also draws on the advantages of good extensibility and high performance from existed open source efforts. TNN has been deployed in multiple Apps from Tencent, such as Mobile QQ, Weishi, Pitu, etc. Contributions are welcome to work in collaborative with us and make TNN a better framework.
1. 环境(environment)
RunTime DEVICE: METAL
2. Github版本
commit(optional):
3. 编译方式(compile method) CMake完整编译参数(full cmake arguments)为默认参数
5. 详细描述bug 情况 (Describe the bug) 在进行文本分类任务时,我这里将onnxruntime推理后端修改为了TNN,希望可以减少一些内存占用。但是将onnxruntime修改为TNN后,内存占用不降反增了。我这里模型大小约为3M,使用onnxruntime时,内存占用均值为22M,峰值为28M,但是切换为TNN后,内存占用稳定在30M,几乎看不到波动,请问这可能是什么原因导致的呢?能提供一些意见吗?