Closed lnSong closed 1 year ago
Hi, what batchsize
did you use? Is this the speed of Lite-Mono-8m
? did you evaluate the speed of Monodepth2 with the same code?
Batchsize = 1, in the code ,it is dummy_input = torch.rand(1, 3,192,640) This is the speed of Lite-Mono. The speed of Monodepth2 is 6ms.
If you check the graphs of speed evaluation in this repo, you could find that when batchsize=1
, Monodepth2 is 7.1ms, and Lite-Mono is 9.1ms.
What CUDA version are you using? Different CUDA versions might have different performances. Also some additional points:
eval
mode.cudnn.benchmark = False
and observe if it affects the speed.I have CUDA Version: 11.4. Thank you very much for your answer, I will try again what you said.
I am now closing this issue. If you have more questions please feel free to reopen this issue or create a new one.
I tested the speed of monodepth2, R-MSFM3, and Lite-Mono on the RTX 3090. Why was Lite-Mono the slowest? monodepth2 is 4.12s, R-MSFM3 is 7.25s,Lite-Mono is 9.44s
I need more information. Do you use the same code as posted in this issue? What batchsize do you use? Have you tried the points that I told you, i.e. set the model to eval mode. set cudnn.benchmark = False and observe if it affects the speed. try different batchsizes.
Thank you very much! When i increase the batchsize, the result will be better
Yes, the batchsize can affect the result. Also, your code for speed evaluation is different from mine. In my evaluation code there are lines to compute the inference time. Please see this. Then, t2-t1
is the time for a batch. You need to do a warmup and then cumulate all the batches.
If you have further questions feel free to contact me. Good luck!
I am running the following code on Ubuntu, under TITAN V, and I get an inference speed of 13.2ms, which is much different from the results in your paper, is this due to the code or the hardware?