Open rafale77 opened 3 years ago
Thank you. I have replaced the YoloV4l-mish in my implementation with YoloV4-P5. The improvement in accuracy is very noticeable in my application: Live streaming from a camera looking at the score % of cars, they are much higher even under much more challenging lighting conditions. I can almost answer my own question: Knowing that I run at fixed FPS instead of loading my CPU/GPU to the maximum, I am looking at CPU/GPU(RTX3070) load. The increase in GPU load Vs. CSP is ~10%m CPU is increased by 15%. I am puzzled by how heavily loaded the CPU is in general. No significant increase in memory usage. I have since optimized the implementation slightly by for example swapping color channels after passing the image tensor to the GPU instead of doing it before, and reducing the input image size. Any reason why the enlargement of the model run in the GPU increases the CPU load this much?
I also had to disable the "auto" option in the letterbox function as it was occasionally yielding tensor size errors.
I trained with my own data set but it was very very very slow. I used the YoloV4-P5 model, input size: 1376 , GPU:RTX-2080TI x 10, batch size:20. whether it is normal or not ?
First of all, congratulations on publishing such fascinating and fantastic work.
I am currently using your YoloV4 pytorch u5 large pretrained model in my home automation setup.