Closed RQ-Wu closed 8 months ago
I training the first stage with 8*A800 80G. However, the max batch size can only be set to 1 on each single GPU. Is that normal?
It’s normal.
I training the first stage with 8*A800 80G. However, the max batch size can only be set to 1 on each single GPU. Is that normal?