Closed deep404 closed 12 months ago
I tried to minimize the batch to 1 and train locally where I have a 3070 8Gb, but without any success
I am afraid that 8GB might be a bit too low. For reference, we used two RTX8000 GPUs which have 48GB of VRAM each to run a batch size of 10. You can find the checkpoints for the model for inference in our release.
I tried to minimize the batch to 1 and train locally where I have a 3070 8Gb, but without any success