Open developer0hye opened 2 years ago
Can you explain more what do you mean?
@khoshsirat
Does spatial pool function mean Channel-only Self-Attention?
Does channel pool function mean Spatail-only Self-Attention?
OK, I see it now:
The spatial_pool
function should be renamed to channel_pool
and the channel_pool
function should be renamed to spatial_pool
.
I have found another discrepancy too:
In the channel_pool
function (which should be renamed to spatial_pool
), softmax
is called after matmul
. But in the paper, in the Spatial-only Self-attention block, softmax
is used before matmul
.
@khoshsirat You are right The location of softmax operation in channel_pool function is different with paper explanation. What's going on? Which one is correct?
Hi guys, I have created a gist to compare this implementation against External-Attention-pytorch's. Through the simple test case, I found that the outputs are different with kaiming init.
Any idea why?
Thanks for sharing your work!
https://github.com/DeLightCMU/PSA/blob/588b370d9f240d38832061a70c275cb6eb81232e/semantic-segmentation/network/PSA.py#L64-L95
It seems that spatial_pool function is the same with Channel-only self attention module.