issues
search
howardyclo
/
papernotes
My personal notes and surveys on DL, CV and NLP papers.
128
stars
6
forks
source link
When and Why are Pre-trained Word Embeddings Useful for Neural Machine Translation?
#13
Open
howardyclo
opened
6 years ago
howardyclo
commented
6 years ago
Metadata
Authors: Ye Qi, Devendra Singh Sachan, Matthieu Felix, Sarguna Janani Padmanabhan and Graham Neubig
Organization: Language Technologies Institute, Carnegie Mellon University
Release Date: 2018 on Arxiv
Link:
https://arxiv.org/pdf/1804.06323.pdf
howardyclo
commented
6 years ago
Summary
Pre-training the word embeddings in the source and/or target languages helps to increase BLEU scores.
Pre-training source language embeddings gains much improvement, indicating that better encoding of the source sentence is important.
Word embeddings are most effective, where there is very little training data but not so little that the system cannot be trained at all.
The gain from pre-training of embeddings may be larger when the source and target languages are more similar.
A priori alignment of embeddings may not be necessary in bilingual scenarios, but is helpful in multi-lingual training scenarios.
Metadata