Closed MarouanGit closed 5 years ago
Your latency observations are correct for yolov2 on XDNNv2.
Yolo typically takes a 608x608 input, and hence incurs large activations, and a heavy data movement penalties. The v2 architecture stalls the systolic array while partial activations are moved off chip.
XDNNv3 will alleviate this, and we should improve to around 25-30 ms.
Stay tuned.
As per accuracy. The yolov2 model we have when ran on FPGA achieved 15.4% mAP as opposed to the published 21.4%. This on ms coco. We did not train our model very long, and we remove leaky relu.
Hi,
I reopen this thread in order to see if the XDNNv3 has been released and if you had the opportunity to test YoloV3 on it ? How is the inference on this new architecture ?
Hi,
I'm running the ML-Suite and YoloV2 retrained on an Amazon Instance F1 with the Xilinx Alveo AMI. I'm obtaining an average inference of 80ms/image with a quiet low accuracy (compared to 30ms/image on a GPU).
My goal is to understand why do we have this huge gap in inference time and also :
Attached the output of the detection on a AWS F1 instance shell_results_xilinx.txt. Any help would be appreciated !
Many thanks, Cheers.