joeyearsley / efficient_densenet_tensorflow

A memory efficient implementation of densenet.
82 stars 18 forks source link

how is the inference time #2

Closed chenxinhua closed 5 years ago

chenxinhua commented 5 years ago

can this efficient denseNet implementation reduce inference time?

joeyearsley commented 5 years ago

No it can not reduce inference times. In fact reducing a densenets inference time is difficult due to all of the concatenation of layers. If you wish to reduce inference time you could freeze the graph and remove excess computation nodes (e.g gradients) and then the only way to get faster is to make a custom c++ op for the denseblocks or entire graph to save on the latency of starting a new op for every layer.