harvardnlp / seq2seq-attn

Sequence-to-sequence model with LSTM encoder/decoders and attention
http://nlp.seas.harvard.edu/code
MIT License
1.26k stars 278 forks source link

Feature/guidedalignment #52

Closed jungikim closed 8 years ago

jungikim commented 8 years ago

This request adds the implementation of the guided alignment as described in Guided Alignment Training for Topic-Aware Neural Machine Translation (Chen et al. 2016)

In summary:

yoonkim commented 8 years ago

Cool! Also relevant: http://arxiv.org/pdf/1609.04186.pdf (the above work claims that cross entropy does slightly better than MSE for training the attention part of the model)