Open andafonin opened 1 year ago
Hi, I wonder how you managed to fit batch size 24 of your 48 GB GPU. Did you use .half() conversion for model and training data during training?
Also to add to this would running 16 bit precision training speed up inference?
Hi, I wonder how you managed to fit batch size 24 of your 48 GB GPU. Did you use .half() conversion for model and training data during training?