Closed socome closed 2 years ago
Thank you guys, I encountered the same situation too. The model accepts a image and output something like segmentation mask. When test the inference time use a loop with same input, performance on some frames are much slow and this occurs regularly. The python lib is built from the master branch serveral days ago follow the dockerfile build instructions based on pytorch:21.10, tested using python call with fp16 in a t4 gpu.
Make sure you are benchmarking with the proper settings. Things like synchronizes and using the cudnn benchmark settings as as well as determinism effect your results. Here is the script we maintain for benchmarking models https://github.com/NVIDIA/Torch-TensorRT/blob/master/examples/benchmark/py/perf_run.py
This issue has not seen activity for 90 days, Remove stale label or comment or this will be closed in 10 days
❓ Question
Thank you for this nice project, I successfully converted my model, which feeds multispectral images, using Torch-TensorRT as below.
Then, I tested the inference time of the model. I found that sometimes it is too slow as below.
How can i solve this problem..? Performance(Miss-rate) of converted model is the same as performance of original model.
Environment
conda
,pip
,libtorch
, source): condaAdditional context