Open NingJinzhong opened 4 months ago
We train on multiple L40 or A40 GPUs (48GB each) but find that a single 48GB GPU is sufficient to train the model, given enough time. We have not tested with lower VRAM GPUs, but lower batch sizes may be feasible.
Hi, could you let me know how much VRAM is needed to train your model? Thanks!