Closed twmht closed 4 years ago
is this the result (https://github.com/ryujaehun/pytorch-gpu-benchmark/blob/master/fig/new_2080ti/GeForce_RTX_2080_Ti_1_gpus__half_model_inference.png) of running resnet with batch size=12?
I just found out that my inference time with 2080_Ti is 15ms with batch size=12, which is different with yours.
any idea?
Maybe the difference is caused by cudnn. The 2080ti has tensorcore and uses tensorcore in some sizes such as 16x16x16. However, in pytorch, this cannot be determined, and cudnn decides, so I think it is.
is this the result (https://github.com/ryujaehun/pytorch-gpu-benchmark/blob/master/fig/new_2080ti/GeForce_RTX_2080_Ti_1_gpus__half_model_inference.png) of running resnet with batch size=12?
I just found out that my inference time with 2080_Ti is 15ms with batch size=12, which is different with yours.
any idea?