POSTECH-CVLab / FastPointTransformer

Official source code of Fast Point Transformer, CVPR 2022
MIT License
269 stars 41 forks source link

[Question] hardware used and training time #1

Closed QuanticDisaster closed 2 years ago

QuanticDisaster commented 2 years ago

Hi

The paper mentions an inference time extremely reduced compared to the original PointTransformer, but I was also curious about the time it took to train the model. What GPUs did you use and how much time did it take to train compared to PointTransformer ?

Thx a lot

chrockey commented 2 years ago

Hi!

Thanks for your interest in our work 😄 . We train Fast Point Transformer on S3DIS with a single A100-80GB GPU (total 80GB) for 30k iterations with batch size 8, and it took 20 hours. For Point Transformer, we train the model with four RTX 3090 GPUs (total 96GB) for ~40k iterations with batch size 16, and it took ~16 hours. For latency, we measure the inference time of both Fast Point Transformer and Point Transformer on the same GPU, RTX 3090.

FYI, the reason why we did not use multi-GPU training is that we found a weird performance drop when we use multi-GPU DDP training with PyTorch Lightning and MinkowskiEngine. This weird problem happened with both MinkowskiNet and Fast Point Transformer. Therefore, we decided not to use it for our work.

QuanticDisaster commented 2 years ago

Thanks for your detailled answer !