Contrastive and non-contrastive self-supervised learning recover global and local spectral embedding methods

R Balestriero, Y LeCun - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Abstract Self-Supervised Learning (SSL) surmises that inputs and pairwise positive
relationships are enough to learn meaningful representations. Although SSL has recently …

Understanding self-predictive learning for reinforcement learning

Y Tang, ZD Guo, PH Richemond… - International …, 2023 - proceedings.mlr.press
We study the learning dynamics of self-predictive learning for reinforcement learning, a
family of algorithms that learn representations by minimizing the prediction error of their own …

The mechanism of prediction head in non-contrastive self-supervised learning

Z Wen, Y Li - Advances in Neural Information Processing …, 2022 - proceedings.neurips.cc
The surprising discovery of the BYOL method shows the negative samples can be replaced
by adding the prediction head to the network. It is mysterious why even when there exist …

The power of contrast for feature learning: A theoretical analysis

W Ji, Z Deng, R Nakada, J Zou, L Zhang - Journal of Machine Learning …, 2023 - jmlr.org
Contrastive learning has achieved state-of-the-art performance in various self-supervised
learning tasks and even outperforms its supervised counterpart. Despite its empirical …

Contrasting the landscape of contrastive and non-contrastive learning

A Pokle, J Tian, Y Li, A Risteski - arXiv preprint arXiv:2203.15702, 2022 - arxiv.org
A lot of recent advances in unsupervised feature learning are based on designing features
which are invariant under semantic data augmentations. A common way to do this is …

Masked prediction: A parameter identifiability view

B Liu, DJ Hsu, P Ravikumar… - Advances in Neural …, 2022 - proceedings.neurips.cc
The vast majority of work in self-supervised learning have focused on assessing recovered
features by a chosen set of downstream tasks. While there are several commonly used …

Semppl: Predicting pseudo-labels for better contrastive representations

M Bošnjak, PH Richemond, N Tomasev, F Strub… - arXiv preprint arXiv …, 2023 - arxiv.org
Learning from large amounts of unsupervised data and a small amount of supervision is an
important open problem in computer vision. We propose a new semi-supervised learning …

Random teachers are good teachers

F Sarnthein, G Bachmann… - International …, 2023 - proceedings.mlr.press
In this work, we investigate the implicit regularization induced by teacher-student learning
dynamics in self-distillation. To isolate its effect, we describe a simple experiment where we …

The edge of orthogonality: A simple view of what makes byol tick

PH Richemond, A Tam, Y Tang… - International …, 2023 - proceedings.mlr.press
Self-predictive unsupervised learning methods such as BYOL or SimSIAM have shown
impressive results, and counter-intuitively, do not collapse to trivial representations. In this …

Rethinking evaluation protocols of visual representations learned via self-supervised learning

JH Lee, D Yoon, BM Ji, K Kim, S Hwang - arXiv preprint arXiv:2304.03456, 2023 - arxiv.org
Linear probing (LP)(and $ k $-NN) on the upstream dataset with labels (eg, ImageNet) and
transfer learning (TL) to various downstream datasets are commonly employed to evaluate …