flrngel / understanding-ai

personal repository
36 stars 6 forks source link

Directional Self-Attention Network for RNN/CNN-Free Language Understanding #3

Open flrngel opened 6 years ago

flrngel commented 6 years ago

https://arxiv.org/abs/1709.04696

this is primitive version of Bi-BloSAN(see more from #2)

1. Introduction

Features of DiSAN

w.r.t means with regard to

3. Background

Self-Attention: using same x as i and j

3.1 Multi-dimensional Attention

3.3 Directional Self-Attention

  1. x->b
  2. token2token
  3. process partly
  4. masking

-inf makes softmax zero