ganler / ResearchReading

General system research material (not limited to paper) reading notes.
GNU General Public License v3.0
20 stars 1 forks source link

MLSys'21 | sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data #48

Closed ganler closed 3 years ago

ganler commented 3 years ago

Short paper. A new approach for model specialization... https://rise.cs.berkeley.edu/wp-content/uploads/2020/01/sensAI_2_pager.pdf https://github.com/GuanhuaWang/sensAI

ganler commented 3 years ago

Background

Class-specific neuron analysis

e.g., decouple a 10-way NN to 10 binary classifiers.

So I think this only works for classification-based NN architecture;

Pruning

One-Vs-All (OVA) reduction

OVA machine learning model reduction is a general approach that reduces a multi-class learning problem into a bunch of simpler problems solvable with binary classifiers.

Unlike unit ablation, OVA is to train multiple binary classifiers with pre-defined structures. (not converting a big model into small binary ones).

Big picture of sensAI

image

  1. Get N binary classifiers from an N-way big model via pruning (one-shot / iterative);
  2. Retrain the binary classifier to regain accuracy;
  3. Combining results back to N-way prediction: Add a soft-max layer; (I think this is not necessary and might be problematic... since different activation numeric range might differ by class...)

Evaluation

Dataset: CIFAR10 Well, it is not enough but it is fine as preliminary results.

Looks good! image