howardyclo / papernotes

My personal notes and surveys on DL, CV and NLP papers.
128 stars 6 forks source link

When and Why are Pre-trained Word Embeddings Useful for Neural Machine Translation? #13

Open howardyclo opened 6 years ago

howardyclo commented 6 years ago

Metadata

howardyclo commented 6 years ago

Summary

  1. Pre-training the word embeddings in the source and/or target languages helps to increase BLEU scores.
  2. Pre-training source language embeddings gains much improvement, indicating that better encoding of the source sentence is important.
  3. Word embeddings are most effective, where there is very little training data but not so little that the system cannot be trained at all.
  4. The gain from pre-training of embeddings may be larger when the source and target languages are more similar.
  5. A priori alignment of embeddings may not be necessary in bilingual scenarios, but is helpful in multi-lingual training scenarios.