Closed yikedouer closed 3 years ago
请问为什么使用了bilinear attention? word embedding已经通过non-linear transformation与char embedding做到了维度对齐,是出于什么考虑使用了bilinear attention呢?
Hi,
I would like to recommend you to read the original paper, in which I have already shown my motivation to use bilinear attention.
Wei
请问为什么使用了bilinear attention? word embedding已经通过non-linear transformation与char embedding做到了维度对齐,是出于什么考虑使用了bilinear attention呢?