Closed wsy-yjys closed 1 year ago
Hi @wsy-yjys,
Thank You for your interest in our work. The inference time you are referring to is measured on a local server with NVIDIA A100 GPU & Intel(R) Xeon(R) Gold 5317 CPU @ 3.00GHz using PyTorch. In this setting the major bottleneck is the channel attention in SDTA module. Also, in w/o Adaptive Branching
we replace the adaptive branching component with standard depth wise convolution of corresponding size. All this might be the reason of nearly equal inference time in both cases. Please let me know if you have any questions or have any additional comments/insights.
Best Regards, Muhammad Maaz
tkank you for your reply!I got it.
In the table 7. I'm curious why the delay is the same with the base model when there is no adaptive branching? Looking forward to your reply~