Open joihn opened 1 year ago
Hi @joihn , thanks for your suggestion. We can have a discussion about the expected usage of auto batch size
and how to implement it in this issue.
Could you first introduce your idea about them?
Reference
Hi @zhouzaida , I'm participating in the OpenMMLab Code Camp task. Currently, I have a rough implementation idea, referring to toma's approach:
However, I have two questions regarding how to implement this functionality in mmengine:
These are my current questions. If there are any misunderstandings in my description of the task, please feel free to point them out.
What is the feature?
When deploying trainning everyday different machine having different GPU, it's tedious to re-tune the batch size manually (trying to maximize it without the training crashing due to
out of memory
)It would be cool to have an "auto batch size" like in yolov5 https://github.com/ultralytics/yolov5
I could implement it myself if someone gives me architectural adivce on where it wold best be implemented
Any other context?
No response