Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)
First of all, thank you for your awesome project!
There was a problem while learning the rec model, so I ask for your help.
When learning with a very large data set, a bottleneck occurs when loading data in 1epoch.
The image below is a graph of GPU utilization when learning 1epoch.
It can be seen that the GPU is used normally when evaluating, but the GPU utilization decreases again when training.
However, after 1epoch, all data is loaded into memory and the GPU operates normally.
If the dataset is small, it is good at using GPU from the beginning.
I want to resolve the bottleneck. Can you solve the bottleneck?
Finally, I attach the contents of my yaml file.
Thanks!
First of all, thank you for your awesome project! There was a problem while learning the rec model, so I ask for your help.
When learning with a very large data set, a bottleneck occurs when loading data in 1epoch. The image below is a graph of GPU utilization when learning 1epoch. It can be seen that the GPU is used normally when evaluating, but the GPU utilization decreases again when training. However, after 1epoch, all data is loaded into memory and the GPU operates normally. If the dataset is small, it is good at using GPU from the beginning.
I want to resolve the bottleneck. Can you solve the bottleneck?
Finally, I attach the contents of my yaml file. Thanks!