Closed iMurphL closed 3 years ago
i would also be pleasure to list some hint in english.
按照以下顺序进行排查:
Yeah I figured it out with your helps. There were some processes using other GPUs which unexpectedly influenced my tests. And the timing module between python and cpp is different. I rewrite the test code with TorchScript and the time is also around 20ms. Thank you for your helps.
Hi there! Thank you for your excellent codes and it helps me a lot. I trained a network with pytorch and deployed it with tensorRT successfully. But the infer time (do NOT include pre/post process) got longer compared to inferring in torch. While converted to INT8 the model is getting faster but not enough. Is that normal? Maybe there is something I missed while deploying the model. I have no idea about it and can you hint me with any ideas? GPU: GTX1080Ti/CUDA10.0 Model: DeeplabV3Plus with backbone ResNet50 pytorch1.6 infer time 15ms tensorrt infer time 22ms/FP32, 13ms/INT8