Closed DuZzzs closed 4 years ago
The inference time is measured on an internal platform based on trt. The INT8 quantization training will give you an INT8 model, but the inference software stack is beyond the current scope of SimpletDet.
On Wed, Dec 25, 2019 at 12:41 PM DuZzzs notifications@github.com wrote:
Thank you for your work. When I reproduce the int8 quantified example, how do I get the inference time of the int8 quantized model? I ran detection_test.py and did not observe speedups. Run a file? Does anyone know? Thanks.
This table: ( https://github.com/TuSimple/simpledet/tree/master/config/int8) dtype time(ms) minival mAP fp32 260 35.7 int8 100 35.8 my environment: gpu : 2060 tensorrt 7.0
— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/TuSimple/simpledet/issues/283?email_source=notifications&email_token=ABGODH4NIMYY74E66BSCGV3Q2LP75A5CNFSM4J7CKGGKYY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4ICSD4IQ, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABGODHYR2QH5RTXY2WNMRRDQ2LP75ANCNFSM4J7CKGGA .
I see. Thank you.
Thank you for your work. When I reproduce the int8 quantified example, how do I get the inference time of the int8 quantized model? I ran detection_test.py and did not observe speedups. Run a file? Does anyone know? Thanks.
my environment: gpu : 2060 tensorrt 7.0