Closed yassouali closed 4 years ago
Please refer to here for all details.
torch.cuda.max_memory_allocated()
. This function always returns the maximum GPU Memory during its lifetime, so we have to organize all modules we want to count in the increasing order of memories, make sure that the result is the actual result of the current module.MatMul
,AdaptiveAvgPool
.torch.cuda.synchronize()
before computing time and test it several rounds to get the average time.torch.autograd.profiler.profile()
. I hope it's clear for you.
Thank you very much for your response.
First of all thank you for this wonderful work, and for providing the implementation.
Feel free to close this issue if this is not the place to ask such questions. I really liked your computation and memory statistics, I was wandering what tool you've used to get such statistics.
Thank you very much