Closed styfeng closed 1 year ago
Update: i just tried using an A40 (40GB of VRAM) and it trains fine with batch size up to 64. it seems the model by default consumes around 26gb of VRAM or more (I believe), so distributing across 24GB GPUs doesn't work
See title. This seems a bit unreasonable to me. I wonder if it's an issue with the script because I doubt 96gb of vram would not be enough to train on batch size of 1 (images are dimensions 256x256)... if anybody has gotten the training working, let me know!