Closed harrylal closed 9 months ago
@harrylal Thank you very much for your comments, I will try to find out the reason soon. Besides, I would also be very grateful if you could give your feedback when finding the cause.
@chequanghuy Thank you for your quick response. I have conducted a thorough model profiling, and it appears that the encoder layer with 131 kernels may be contributing to the performance issue on CPU. I would highly value your insights on this.
A big shoutout to the amazing folks who brought the TwinLiteNet model to life.I'm genuinely impressed by what you've accomplished here. Thanks a million for your outstanding contribution! ππ.
I'd like to discuss a matter where I could use your insights. I've been running the TwinLiteNet model on an Intel i9 CPU and it's been delivering about 0.5 frames per second (fps). In comparison, I've observed that YOLOv8n, despite having significantly more parameters, achieves around 15 fps on the same CPU.
Current Behavior:
The TwinLiteNet model performs at approximately 0.5 fps on the specified CPU configuration.
Expected Behavior:
I'm seeking insights on what could be possible bottlenecks, tips, or tweaks that can potentially improve the TwinLiteNet model's CPU inference speed to achieve better performance, even in comparison to YOLOv8n with significantly more parameters.
Steps to Reproduce:
Modify and run test_image.py as below for inference on CPU and FPS logging. Reference - https://github.com/chequanghuy/TwinLiteNet/issues/2#issuecomment-1667666914