kaiyuyue / cgnl-network.pytorch

Compact Generalized Non-local Network (NIPS 2018)
https://arxiv.org/abs/1810.13125
MIT License
259 stars 41 forks source link

Normalization #1

Closed shaibagon closed 6 years ago

shaibagon commented 6 years ago

Hi, I find your work very interesting. However I have two questions regarding normalization:

  1. In the original non-local neural networks work, the product of phi and theta is normalized BEFORE it multiplies g to produce the output (in their work it is done using a softmax layer). I do not see any such normalization in your work - why?

  2. Your Taylor expansion is based on the assumption that both theta and phi are of unit L2 norm. I do not see this enforced in your code - what have I missed?

Thanks,

kaiyuyue commented 6 years ago

Hi Shai,

Thanks for the interests.

shaibagon commented 6 years ago

@KaiyuYue thank you for the detailed answer.