Open HsunGong opened 1 year ago
Hello
Related to #15573
@HsunGong could you pls share your use case when these various batch sizes could be used? probably if you have a heterogenous set of GPU cards?
For sure, we have 2080ti, 3090, a10, a40 gpu cards, and we want to put dynamic batch size according to the gpu. We've made it using pytorch-ddp by customizing samplers for each gpu.
Description & Motivation
Support like:
GPU1: batch-size=24 GPU2: batch-size=12 GPU3: batch-size=16 GPU4: batch-size=24
As different batch has different length
See
https://github.com/microsoft/DeepSpeed/issues/1051
https://github.com/facebookresearch/fairseq/blob/b5a039c292facba9c73f59ff34621ec131d82341/fairseq/data/data_utils.py#L282
Pitch
No response
Alternatives
No response
Additional context
No response
cc @borda