a1da4 / paper-survey

Summary of machine learning papers
32 stars 0 forks source link

Reading: Biased Random Walk based Social Regularization for Word Embeddings #227

Open a1da4 opened 2 years ago

a1da4 commented 2 years ago

0. Paper

@inproceedings{ijcai2018-634, title = {Biased Random Walk based Social Regularization for Word Embeddings}, author = {Ziqian Zeng and Xin Liu and Yangqiu Song}, booktitle = {Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, {IJCAI-18}}, publisher = {International Joint Conferences on Artificial Intelligence Organization},
pages = {4560--4566}, year = {2018}, month = {7}, doi = {10.24963/ijcai.2018/634}, url = {https://doi.org/10.24963/ijcai.2018/634}, }

1. What is it?

They updated socialized word representation with a biased random walk.

2. What is amazing compared to previous works?

Precious works apply to one-hop neighbors (friends, followers), so there are texts input by users with more friends than texts input by users with fewer friends. To avoid this limitation, their method obtain neighbors with a biased random walk.

3. Where is the key to technologies and techniques?

3.1 Socialized Word Embeddings (2017)

スクリーンショット 2021-12-31 14 21 16

  1. use all documents, train CBOW-based word embeddings スクリーンショット 2021-12-31 14 23 37

  2. to obtain user embeddings, they train node2vec (one-hop neighbors are similar). スクリーンショット 2021-12-31 14 23 56

  3. finally, the training objective function is defined as below: スクリーンショット 2021-12-31 14 24 15

From this model, a user-specific word vector (user i, word j) is defined. スクリーンショット 2021-12-31 14 27 41

3.2 Biased Second-Order Random Walk

The above model (3.1) only used one-hop neighbors, number of friends of users has a direct impact on training. To avoid this limitation, they applied a biased random walk (second-order random walk). スクリーンショット 2021-12-31 14 32 58 スクリーンショット 2021-12-31 14 33 20

A bias term is defined as follows.

4. How did evaluate it?

スクリーンショット 2021-12-31 14 36 24 スクリーンショット 2021-12-31 14 36 45

5. Is there a discussion?

6. Which paper should read next?