Open thisistejaspandey opened 3 months ago
Hi, in the RTDETR paper, the network was trained for a batch size of 4 for 4 GPUs.
Why did you choose such a small batch size and would you expect better results with a larger batch?
Many thanks!
p40
bs
Hi, in the RTDETR paper, the network was trained for a batch size of 4 for 4 GPUs.
Why did you choose such a small batch size and would you expect better results with a larger batch?
Many thanks!