Open tensorflowt opened 4 years ago
May I ask whether the current project supports INT8 quantization? If so, how? Currently onlyFT16, FT32 quantification is supported, right?
Correct, but I think you can do PTQ using TensorRT calibration. Not implemented in this repo and not planned to do so in short phase.
May I ask whether the current project supports INT8 quantization? If so, how? Currently onlyFT16, FT32 quantification is supported, right?