Wechat ID: NeuralTalk
关注模型压缩、低比特量化、移动端推理加速优化、部署
A curated list of awesome A.I. & Embedded/Mobile-devices resources, tools and more.
Looking for contributors. Submit a pull request if you have something to add :)
Please check the contribution guidelines for info on formatting and writing pull requests.
Qualcomm Adreno GPU Performance as below:
[1606.05316] Learning Infinite-Layer Networks: Without the Kernel Trick
[1608.02893] Syntactically Informed Text Compression with Recurrent Neural Networks
[1608.05148] Full Resolution Image Compression with Recurrent Neural Networks
[1707.09422] Hyperprofile-based Computation Offloading for Mobile Edge Networks
[1707.09855] Convolution with Logarithmic Filter Groups for Efficient Shallow CNN
[1707.09597] ScanNet: A Fast and Dense Scanning Framework for Metastatic Breast Cancer Detection from Whole-Slide Images
[1604.08772] Towards Conceptual Compression
ARM-software/ComputeLibrary: The ARM Computer Vision and Machine Learning library is a set of functions optimised for both ARM CPUs and GPUs using SIMD technologies Intro
mil-tokyo/webdnn: Fastest DNN Execution Framework on Web Browser
jiaxiang-wu/quantized-cnn: An efficient framework for convolutional neural networks
naibaf7/libdnn: Greentea LibDNN - a universal convolution implementation supporting CUDA and OpenCL
dmlc/nnvm-fusion: Kernel Fusion and Runtime Compilation Based on NNVM
Model convertor. More convertors please refer deep-learning-model-convertor
This part contains related course, guides and tutorials.
Deep learning systems: UW course schedule(focused on systems design, not learning)
Efficient Convolutional Neural Network Inference on Mobile GPUs
Tutorial on Hardware Architectures for Deep Neural Networks | MIT MICRO-50
Creating insanely fast image classifiers with MobileNet in TensorFlow | HACKERNOON
Building Cross-Platform CUDA Applications with CMake | NVIDIA
2017-08-07
2017-07-24