Closed Lycan1003 closed 3 years ago
The reason why we use old_weight
as parameter is that we want to update weight parameter in high precision(fp32) to increase accuracy and speed up training convergence process. Of course, we can handle bias
in the same way. However, we think int32 precision is accurate enough for bias
and the bias updating on this precision will not influence accuracy and training process.
The reason why we use
old_weight
as parameter is that we want to update weight parameter in high precision(fp32) to increase accuracy and speed up training convergence process. Of course, we can handlebias
in the same way. However, we think int32 precision is accurate enough forbias
and the bias updating on this precision will not influence accuracy and training process.
Thanks for your reply!
compressor.py line 467
In this class, it uses 'old_weight' as the parameter and 'weight' as the buffer. I wonder to know that why don't do the same operation to the bias? In quantizer.py line
It seems to quantize aware bias directly.