a1da4 / paper-survey

Summary of machine learning papers
32 stars 0 forks source link

Reading: On the Sentence Embeddings from Pre-trained Language Models #258

Open a1da4 opened 1 year ago

a1da4 commented 1 year ago

0. Paper

1. What is it?

They propose a method to convert sentence embeddings into an isotropic Gaussian space to improve sentence representation.

2. What is amazing compared to previous works?

Their method improves the performance of sentence textual similarity tasks.

3. Where is the key to technologies and techniques?

Convert vector space (anisotropy, biased) into Gaussian space (isotropy, unbiased)

スクリーンショット 2023-05-03 11 24 04

During training, they optimize the following function as follows:

スクリーンショット 2023-05-03 11 24 44

They performed spatial transformations by NN. The only parameters that are adjusted during training are those of the NN and never change the parameters of the BERT (computational resource friendly).

4. How did evaluate it?

スクリーンショット 2023-05-03 11 29 32

From Table 2, their method constantly outperforms baselines.

5. Is there a discussion?

vs Standard Normalization, and all-but-the-top (NATSV in this paper) #51

スクリーンショット 2023-05-03 11 31 39

Table 5 shows that their method outperforms these previous works. Thay mention that

We argue that NATSV can help eliminate anisotropy but it may also discard some useful information contained in the nulled vectors. On the contrary, our method directly learns an invertible mapping to isotropic latent space without discarding any information.

(→ What is "useful information"? )

6. Which paper should read next?