Closed NorbertZheng closed 1 year ago
Self-Supervised Learning.
Learning classification with Unlabeled Data. de Sa NIPS’93, University of Rochester. 1993 NIPS, Over 200 Citations. Self-Supervised Learning, Unsupervised Learning, Multimodal.
A piecewise-linear classifier in a 2-Dimensional input space.
Self-Supervised Learning.
For example, hearing “mooing” and seeing cows tend to occur together.
So, although the sight of a cow does not come with an internal homuncular “cow” label it does co-occur with an instance of a “moo”.
Network for Learning the Labels of the Codebook Vectors.
We should note that $H=WX$ computes the similarity between input $X$ and codebook vectors $W$. And implicit labeling weights map each codebook vectors $W$ to their corresponding labels.
One way to make use of the cross-modality structure is to derive labels for the codebook vectors. The labels can be learnt with a competitive learning algorithm using a network.
If modality 1 experiences a sensation from its pattern A distribution, modality 2 experiences a sensation from its own pattern A distribution.
The following experiments were all performed using the Peterson and Barney vowel formant data.
The dataset consists of the first and second formants for ten vowels in a /h V d/context from 75 speakers (32 males, 28 females, 15 children) who repeated each vowel twice.
Accuracy (mean percent correct and sample standard deviation over 60 trials and 2 modalities). The heading i-j refers to performance measured after the j-th step during the i-th iteration.
These results were then averaged over 60 runs.
76%–79% accuracies are obtained with 2 different settings.
(It’s the first day of Chinese new year in 2022, I ‘ve just read it very quickly and present it very roughly. For more details, please read the paper. It is amazing that there is a self-supervised learning in the year of 1994.)
Sik-Ho Tang. Review — Learning classification with Unlabeled Data.