Closed jjsr closed 2 years ago
First, I don't know much about the attention, but I'll try to explain to you what I understand This is the source code of attention that I used. It's in Vietnamese so pls use google translate: https://pbcquoc.github.io/vietnamese-ocr/
Thanks again brother sorry for posting at two places I am closing the issue now thanks again
Sir the explanation is awesome. Thanks Can you tell please more about attention- 1) What type of attention it is ? Is it self attention (giving the weighted sum of attention vector at every time step) 2) What is the reason for applying it just after CNN not apply it after RNN 3) Sir in code you have done the following steps regarding attention - a) Permute b) Dense layer c) Mean (dimensionality reduction) d) Repeat vector 4) Layers multiply
It would be great help sir . I can understand dense layer is implementing tanH function but unable to understand the reason behind other operations
May I mail you sir .. Thank you in advance . I was unable to understand the reasons behind the code from quite few time