Closed GauravBh1010tt closed 6 years ago
My code for attention is correct. I got the intended results on the datasets used in the papers. I also contacted the authors of Improved Representation Learning for Question Answer Matching about the dimensions of the attention parameters and it looks like I am on the right track.
I am trying to implement word level attention as described in Teaching Machines to Read and Comprehend and Improved Representation Learning for Question Answer Matching. My code works while I am still not sure whether the dimensions of my attention parameters are correct. This is what it looks like:-
Following previous thread on the same topic - Issue 4962, Issue 1472, I came up I came up with following code snippet