Open severecoder opened 6 months ago
I am moving to quantization aware training to improve the accuracy, to improve the quantized int8 model, is pytorch_quantization the best tool for that?
pytorch_quantizaton will be deprecated, please use AMMO now.
Thank for the response, isn't ammo only limited to LLMs?
I have used PTQ for int8 export from pytorch model and despite attempts at calibration, there is a significant drop in detection accuracy.
I am moving to quantization aware training to improve the accuracy, to improve the quantized int8 model, is pytorch_quantization the best tool for that?
The end result is to have .trt or engine file inferencing at int8 precision with best possible detection metrics.
TIA