Closed Phuoc-Hoan-Le closed 1 year ago
When using distributed training, I see one GPU using 20GB of memory and another GPU using 30GB of memory. Do you guys encounter this weird problem?
When using distributed training, I see one GPU using 20GB of memory and another GPU using 30GB of memory. Do you guys encounter this weird problem?