Understand and improve contrastive learning methods for visual representation: A review

R Liu - arXiv preprint arXiv:2106.03259, 2021 - arxiv.org
Traditional supervised learning methods are hitting a bottleneck because of their
dependency on expensive manually labeled data and their weaknesses such as limited …

Benchmarking representation learning for natural world image collections

G Van Horn, E Cole, S Beery, K Wilber… - Proceedings of the …, 2021 - openaccess.thecvf.com
Recent progress in self-supervised learning has resulted in models that are capable of
extracting rich representations from image collections without requiring any explicit label …

Siamese image modeling for self-supervised vision representation learning

C Tao, X Zhu, W Su, G Huang, B Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
Self-supervised learning (SSL) has delivered superior performance on a variety of
downstream vision tasks. Two main-stream SSL frameworks have been proposed, ie …

Rejuvenating image-gpt as strong visual representation learners

S Ren, Z Wang, H Zhu, J Xiao, A Yuille… - Forty-first International …, 2023 - openreview.net
This paper enhances image-GPT (iGPT), one of the pioneering works that introduce
autoregressive pretraining to predict the next pixels for visual representation learning. Two …

Unsupervised representation learning using convolutional and stacked auto-encoders: a domain and cross-domain feature space analysis

GB Cavallari, LSF Ribeiro… - 2018 31st SIBGRAPI …, 2018 - ieeexplore.ieee.org
A feature learning task involves training models that are capable of inferring good
representations (transformations of the original space) from input data alone. When working …

Revisiting the critical factors of augmentation-invariant representation learning

J Huang, X Kong, X Zhang - European Conference on Computer Vision, 2022 - Springer
We focus on better understanding the critical factors of augmentation-invariant
representation learning. We revisit MoCo v2 and BYOL and try to prove the authenticity of …

Vision mamba: Efficient visual representation learning with bidirectional state space model

L Zhu, B Liao, Q Zhang, X Wang, W Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently the state space models (SSMs) with efficient hardware-aware designs, ie, the
Mamba deep learning model, have shown great potential for long sequence modeling …

Solving inefficiency of self-supervised representation learning

G Wang, K Wang, G Wang… - Proceedings of the …, 2021 - openaccess.thecvf.com
Self-supervised learning (especially contrastive learning) has attracted great interest due to
its huge potential in learning discriminative representations in an unsupervised manner …

Pushing the limits of self-supervised resnets: Can we outperform supervised learning without labels on imagenet?

N Tomasev, I Bica, B McWilliams, L Buesing… - arXiv preprint arXiv …, 2022 - arxiv.org
Despite recent progress made by self-supervised methods in representation learning with
residual networks, they still underperform supervised learning on the ImageNet classification …

Decentralized unsupervised learning of visual representations

Y Wu, Z Wang, D Zeng, M Li, Y Shi, J Hu - arXiv preprint arXiv:2111.10763, 2021 - arxiv.org
Collaborative learning enables distributed clients to learn a shared model for prediction
while keeping the training data local on each client. However, existing collaborative learning …