Closed ClarkChin08 closed 4 years ago
you can download imagenet validation set from kaggle or academictorrents
another question is about 16-bits Quantization for Bias, that means the bias is 16-bits? or the weight should be quantized to 16-bits? if the weight was quantized to 16-bits, the accuracy improved may come from the high precision of weight, right?
weight: 8 bits activation: 8 bits bias: 16 bits
All quantization in this repo is per-tensor based.
Bias quantized to 8 bits also seems to work fine for me.
@bangawayoo well, that's good to know! It probably because I've fixed some minor bugs in set_quant_minmax function recently. I'll update the 8 bits results after I test it.
in file main_cls.py