NorbertZheng / read-papers

My paper reading notes.
MIT License
8 stars 0 forks source link

Sik-Ho Tang | Review -- Learning classification with Unlabeled Data. #117

Closed NorbertZheng closed 1 year ago

NorbertZheng commented 1 year ago

Sik-Ho Tang. Review — Learning classification with Unlabeled Data.

NorbertZheng commented 1 year ago

Overview

image Self-Supervised Learning.

Learning classification with Unlabeled Data. de Sa NIPS’93, University of Rochester. 1993 NIPS, Over 200 Citations. Self-Supervised Learning, Unsupervised Learning, Multimodal.

NorbertZheng commented 1 year ago

Piecewise-Linear Classifier

image A piecewise-linear classifier in a 2-Dimensional input space.

NorbertZheng commented 1 year ago

Self-Supervised Piecewise-Linear Classifier

Cow and “Moo”

image Self-Supervised Learning.

For example, hearing “mooing” and seeing cows tend to occur together.

So, although the sight of a cow does not come with an internal homuncular “cow” label it does co-occur with an instance of a “moo”.

NorbertZheng commented 1 year ago

Self-Supervised Piecewise-Linear Classifier

image Network for Learning the Labels of the Codebook Vectors.

We should note that $H=WX$ computes the similarity between input $X$ and codebook vectors $W$. And implicit labeling weights map each codebook vectors $W$ to their corresponding labels.

One way to make use of the cross-modality structure is to derive labels for the codebook vectors. The labels can be learnt with a competitive learning algorithm using a network.

If modality 1 experiences a sensation from its pattern A distribution, modality 2 experiences a sensation from its own pattern A distribution.

NorbertZheng commented 1 year ago

Experimental Results

The following experiments were all performed using the Peterson and Barney vowel formant data.

The dataset consists of the first and second formants for ten vowels in a /h V d/context from 75 speakers (32 males, 28 females, 15 children) who repeated each vowel twice.

image Accuracy (mean percent correct and sample standard deviation over 60 trials and 2 modalities). The heading i-j refers to performance measured after the j-th step during the i-th iteration.

These results were then averaged over 60 runs.

76%–79% accuracies are obtained with 2 different settings.

(It’s the first day of Chinese new year in 2022, I ‘ve just read it very quickly and present it very roughly. For more details, please read the paper. It is amazing that there is a self-supervised learning in the year of 1994.)

NorbertZheng commented 1 year ago

Reference