Closed james77777778 closed 3 years ago
Yeah, this is a device-specific problem. Depth-wise Conv in YOLOX-nano is not friendly to the device with a low memory access band. Also, it may be caused by the sub-optimization of DWConv in Pytorch. Maybe you can try TorchScript.
Thanks for quick reply!
I think I will stick with onnxruntime for faster inference speed
Thank you for sharing this project!
I encounter a strange problem that YOLOX-nano inference speed is abnormal when I use PyTorch with CPU I tried to benchmark different model on a arm64 computer and I use following script
In my device (ODROID-C4) YOLOX-tiny gets 0.75 FPS YOLOX-nano gets 0.32 FPS (far slower than YOLOX-tiny)
But when I test with onnxruntime YOLOX-tiny gets 1.3 FPS YOLOX-nano gets 3.2 FPS and the result is fine
I can confirm the preprocessing and postprocessing spent similar time and the FPS gap resulted from model inference
maybe this is device-specific problem?! (expensive operation in YOLOX-nano?)