pytorch / TensorRT

PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
https://pytorch.org/TensorRT
BSD 3-Clause "New" or "Revised" License
2.6k stars 350 forks source link

❓ [Question] Model inference is much slower after updating to TensorRT 9.3 #2822

Open demuxin opened 6 months ago

demuxin commented 6 months ago

❓ Question

I have a VIT model for object detection. The model inference speed in the tensort 8.5 environment is 190ms per frame. However when I updated to TensorRT 9.3, Inference slowed down to 250ms per frame.

I acquired the C++ dynamic library by compiling the latest Torch-TensorRT source code.

What might be causing this issue?

Environment

Build information about Torch-TensorRT can be found by turning on debug messages

narendasan commented 6 months ago

We have not tested TRT 9.3 with Torch-TRT so its hard to say why you are seeing this slowdown