Closed bnaman50 closed 5 years ago
Hi Naman,
yes, indeed for a one-layer network w/o relu they are the same except for the scaling. The gradient-backprop is initialized with a 1, while Deconvnet and GB are initialized with the output value of the functions. Does this answer your question?
Cheers, Max
Hey Alber,
I was looking at the analysis results of gradient vs Guided Backprop/DeConvNet methods for simple models. After reading the papers, my impression is that these methods are similar except for how they deal with the non-linearity. So if you have a simple 1-layer deep network with no non-linearity in the network, one would expect to see the exact same results but the results are different. To be precise, the results of guided-backprop/DeConvNet are a scaled version of gradient analysis which is weird.
Here is the code to reproduce the results.
Here is the terminal log.
It would be great if you could explain as to why is this happening.
Thanks, Naman