

Self-Supervised Learning
source link: https://www.tuicool.com/articles/u2Mfem3
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.


Self-Supervision is in the air . Explaining the difference between self-, un-, weakly-, semi-, distantly-, and fully-supervised learning (and of course, RL) just got a 100 times tougher. :) Nevertheless, we are gonna try.
To encode an object (a word, sentence, image, video, audio, …) into a general-enough representation (blobs of numbers), you try to setup learning tasks between parts of it or different views of it (the self).
Given one part (input) of the object,
can you predict / generate the other part (output)?
For example, given a sentence context around a word, can you (learn to) predict the missing word (skip-grams, BERT). Or, modify the view of an object and predict what changed (rotate an image and predict the rotation angle). Because you are simply playing around with the object, these are free lunch tasks — no external labels needed.
But, now that you have (plenty of) auto-generated input-output examples , go ahead and use every hammer from your supervised learning toolkit to learn a great (universal?) representation for the object.
By trying to predict the self-output from the self-input, you end up learning about the intrinsic properties / semantics of the object, which otherwise would have taken a ton of examples to learn from.

Self-supervision losses have been the silent heroes for a while now, across representation learning for multiple domains (as auto-encoders , word embedders , auxiliary losses, … ). A very nice slide deck here . Now, with the ImageNet moment for NLP (ELMo, BERT and others), I guess they’ve made it on their own. The missing gap in the supervision spectrum that everyone (including AGI) has been waiting for.
Understandably, there is flurry of research activity around newer self-supervision tricks , getting SoTA with fewer examples , and mixing various kinds of supervisions (hello NeurIPS !). Till now, the self-supervised methods mostly try to relate the components of an object, taking one part as input, predict the other part. Let’s see how creative the community gets when playing around with the new hammer.
Also, I’m very curious who claims they were the first to do it :)
PS: if you are looking for someone to ‘supervise’ you (weakly, fully, remotely or even co-supervise) to solve some very interesting text, vision and speech problems, feel free to ping me at [email protected] !
Recommend
-
22
Contrastive self-supervised learning techniques are a promising class of methods that build representations by learning to encode what makes two things similar or different. The prophecy that self-supervised...
-
8
This real-world multiplayer game lets you spend other people’s money — really by Mix — 4 months ago...
-
8
Self-supervised Video Object Segmentation by Motion Grouping Self-supervised Video Object Segmentation by Motion Grouping Visual Geometry Group, University of Oxford
-
6
Facebook details self-supervised AI that can segment images and videos Join Transform 2021 this July 12-16. Register fo
-
12
论文地址:https://openaccess.thecvf.com/content/CVPR2021/papers/Hu_Self-Supervised_3D_Mesh_Recons...
-
10
AV-HuBERT (Audio-Visual Hidden Unit BERT) Learning Audio-Visual Speech Representation by Masked Multimodal Cluster Prediction Robust Sel...
-
7
Why self-supervised learning is a medical AI game-changer Skip to main content...
-
3
This article was published as a part of the Data Science Blogathon.
-
7
[Submitted on 24 Apr 2023] A Cookbook of Self-Supervised Learning Download PDF
-
7
[Submitted on 24 May 2023 (v1), last revised 31 May 2023 (this version, v2)] Reverse Engineering Self-Sup...
About Joyk
Aggregate valuable and interesting links.
Joyk means Joy of geeK