@inproceedings{huang-etal-2012-improving,
title = "Improving Word Representations via Global Context and Multiple Word Prototypes",
author = "Huang, Eric and
Socher, Richard and
Manning, Christopher and
Ng, Andrew",
booktitle = "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2012",
address = "Jeju Island, Korea",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/P12-1092",
pages = "873--882",
}
1. What is it?
They used not only information on context words (local information) but also information on the entire document (global information) to learn word vectors considering polysemous words.
2. What is amazing compared to previous works?
They adopted a neural language model to handle local and global information
They created a dataset of the word similarity task that take polysemy into account
3. Where is the key to technologies and techniques?
local score: concatenate ordered word vectors x (xm is a target word) in each sentence
global score: using weighted average word vectors d in the document
final score:
From Reisinger and Mooney (2010), their model can learn multi-prototype word vectors.
4. How did evaluate it?
4.1 Nearest neighbors
Table 2 shows that each prototype describes each sense of target words.
4.2 Similarity tasks
WordSim-353
From Table 3, their model outperformed the previous single-prototype method (C&W).
Word Similarity in Context (new)
From Table 5, their model achieved state-of-the-art performance.
0. Paper
@inproceedings{huang-etal-2012-improving, title = "Improving Word Representations via Global Context and Multiple Word Prototypes", author = "Huang, Eric and Socher, Richard and Manning, Christopher and Ng, Andrew", booktitle = "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = jul, year = "2012", address = "Jeju Island, Korea", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/P12-1092", pages = "873--882", }
1. What is it?
They used not only information on context words (local information) but also information on the entire document (global information) to learn word vectors considering polysemous words.
2. What is amazing compared to previous works?
3. Where is the key to technologies and techniques?
local score: concatenate ordered word vectors x (xm is a target word) in each sentence
global score: using weighted average word vectors d in the document
final score:
From Reisinger and Mooney (2010), their model can learn multi-prototype word vectors.
4. How did evaluate it?
4.1 Nearest neighbors
Table 2 shows that each prototype describes each sense of target words.
4.2 Similarity tasks
WordSim-353 From Table 3, their model outperformed the previous single-prototype method (C&W).
Word Similarity in Context (new) From Table 5, their model achieved state-of-the-art performance.
5. Is there a discussion?
6. Which paper should read next?