Open FrankCAN opened 5 years ago
Hello, do you figure it out this part? I'm also confused about the attn_head() function.
Hello, do you figure it out this part? I'm also confused about the attn_head() function.
I also confused about the attn__head() function. can I learn more about you to figure it out.
Hi, Sorry for me late response, it is the same as sparse GAT.
Please check the paper Neural machine translation by jointly learning to align and translate.
Hi Petar, Many thanks for your code.
For the f_1 to transpose(f_2) operation, can I change the f_2 to f_2_neighbor (BxNxKx1) which means that obtain all the k neighbors first before broadcasting. And then operate f_1 + f_2_neighbor. The total dimension will become to BxNxKx1 directly. Do you think it is the same as sparse GAT? Many thanks for your help.
Best Regards Frank