joapolarbear / dl_notes

1 stars 1 forks source link

HAWQ: Hessian AWare Quantization of Neural Networks With Mixed-Precision #13

Open joapolarbear opened 4 years ago

joapolarbear commented 4 years ago

ICCV 2019

PDF

Contribution

Automatically select the relative quantization precision of each layer, i.e., use fp16, int8 or one bit

  1. not all layers have the same distribution of floating-point values
  2. the network can have significantly different sensitive to the quantization of each layer.

Multi-stage fine-tuning -- define a fine-tuning order

re-train, sort layers in descending order according to the product of \lambda and weight difference.

joapolarbear commented 4 years ago

image