number9473 / nn-algorithm

algorithm for neural network
259 stars 59 forks source link

Pruning Convolutional Neural Networks for Resource Efficient Inference #124

Open joyhuang9473 opened 7 years ago

joyhuang9473 commented 7 years ago

Pruning Convolutional Neural Networks for Resource Efficient Inference

joyhuang9473 commented 7 years ago

Lisandro79/JetsonCaffe: Pruning Neural Networks

https://github.com/Lisandro79/JetsonCaffe/wiki/Pruning-Neural-Networks

joyhuang9473 commented 7 years ago

PerforatedCNNs accelerate convolutional neural networks (CNNs) by skipping evaluation of the convolutional layers in some of the spatial positions.

https://github.com/mfigurnov/perforated-cnn-caffe

joyhuang9473 commented 7 years ago

Add pruning possibilities at inner_product_layer

https://github.com/BVLC/caffe/pull/4294

joyhuang9473 commented 7 years ago

Apply simple pruning on Caffemodel

https://github.com/garion9013/impl-pruning-caffemodel

joyhuang9473 commented 7 years ago

TensorFlow implementation of "Iterative Pruning"

https://github.com/garion9013/impl-pruning-TF

xizi commented 7 years ago

Hi, joyhuang9473, is there a feasible code project for accelerate my vgg19 model which is trained on private dataset?

ghost commented 6 years ago

I'd be interested in anyone has done work on actually detecting when a pruned (simplified) model is less accurate... that is, if I present you with a full model and a simplified (albeit less accurate) model... can we first look at the input and determine if an input is well classified by the less accurate model (i.e. it's "easy" to classify) or that it's in a "difficult zone" where the less accurate model is subject to significantly less accuracy than the full model?

In other words - given an accurate and less accurate pair of models (where one was pruned from the other), can I tag an input to go confidently down the less accurate (although likely faster) pruned model or tag it to go to the full model?