Open serser opened 3 years ago
CUDA : 11.1 TensorRT : 7.2.2-1 GPU : RTX3090
Hi @CarkusL, can you give us the inference time for batch_size =1 of your TensorRT implementation including also the preprocess and postprocess?
Hi @CarkusL, can you give us the inference time for batch_size =1 of your TensorRT implementation including also the preprocess and postprocess?
[09/15/2021-10:38:05] [I] PreProcess Time: 13.3244 ms [09/15/2021-10:38:05] [I] inferenceDuration Time: 13.3018 ms [09/15/2021-10:38:05] [I] PostProcessDuration Time: 7.13283 ms
Hi @CarkusL, can you give us the inference time for batch_size =1 of your TensorRT implementation including also the preprocess and postprocess?
[09/15/2021-10:38:05] [I] PreProcess Time: 13.3244 ms [09/15/2021-10:38:05] [I] inferenceDuration Time: 13.3018 ms [09/15/2021-10:38:05] [I] PostProcessDuration Time: 7.13283 ms
Hi@CarkusL, are you running this with fp32 or with fp16 ?
Hi @CarkusL, can you give us the inference time for batch_size =1 of your TensorRT implementation including also the preprocess and postprocess?
[09/15/2021-10:38:05] [I] PreProcess Time: 13.3244 ms [09/15/2021-10:38:05] [I] inferenceDuration Time: 13.3018 ms [09/15/2021-10:38:05] [I] PostProcessDuration Time: 7.13283 ms
Hi@CarkusL, are you running this with fp32 or with fp16 ?
FP32 on cpu
@CarkusL Thanks for your great work, I wrote a new project based on your code, where computations of pre-process && post-process are done with Cuda, it runs much faster.
Here is the code : https://github.com/Abraham423/CenterPointTensorRT.git
Hi @CarkusL , I am using Tensor 7.2.3.4 on V100. I find the latency is almost twice slower than reported. Could you share us your specific environment settings?