Closed s95huang closed 3 months ago
The first runtime should be slow. It will go through some warming up and memory preallocation. So if you would better like to check the speed for continuous inference, please test the run-time of the second run (in the same python session).
Thank you for your reponse~ Indeed, the second inference becomes much faster, as shown below:
inference time: 1.2322235107421875
inference time: 0.01637864112854004
Closing this issue
Hello, thanks for this work. I am currently testing the mono3d_yolox_576_768.onnx on a single kitti dataset image However, I have observed that the model inference speed is very slow, about 0.9 second per image. Since the model input is quite large, can I ask if you observe similar behavior?
I have set onnxruntime to GPU mode and set
The output on 3090 is: