makrai / toread

Papers I would like to read
0 stars 0 forks source link

nmt (Gehring+ 17, Vaswani+ 17+) #16

Open makrai opened 6 years ago

makrai commented 6 years ago

variants of the LSTM-based Sequence to Sequence with Attention model, particularly Google Neural Machine Translation, were superseded first by a fully convolutional sequence to sequence model and then by the Transformer (Attention is all, Vaswani+ 2017+)