Gpipe: Efficient training of giant neural networks using pipeline parallelism

Y Huang, Y Cheng, A Bapna, O Firat… - Advances in neural …, 2019 - proceedings.neurips.cc
Scaling up deep neural network capacity has been known as an effective approach to
improving model quality for several different machine learning tasks. In many cases …

Autoaugment: Learning augmentation strategies from data

ED Cubuk, B Zoph, D Mane… - Proceedings of the …, 2019 - openaccess.thecvf.com
Data augmentation is an effective technique for improving the accuracy of modern image
classifiers. However, current data augmentation implementations are manually designed. In …

Autoaugment: Learning augmentation policies from data

ED Cubuk, B Zoph, D Mane, V Vasudevan… - arXiv preprint arXiv …, 2018 - arxiv.org
In this paper, we take a closer look at data augmentation for images, and describe a simple
procedure called AutoAugment to search for improved data augmentation policies. Our key …

Do better imagenet models transfer better?

S Kornblith, J Shlens, QV Le - Proceedings of the IEEE/CVF …, 2019 - openaccess.thecvf.com
Transfer learning is a cornerstone of computer vision, yet little work has been done to
evaluate the relationship between architecture and transfer. An implicit hypothesis in …

Re-labeling imagenet: from single to multi-labels, from global to localized labels

S Yun, SJ Oh, B Heo, D Han… - Proceedings of the …, 2021 - openaccess.thecvf.com
ImageNet has been the most popular image classification benchmark, but it is also the one
with a significant level of label noise. Recent studies have shown that many samples contain …

Debiased self-training for semi-supervised learning

B Chen, J Jiang, X Wang, P Wan… - Advances in Neural …, 2022 - proceedings.neurips.cc
Deep neural networks achieve remarkable performances on a wide range of tasks with the
aid of large-scale labeled datasets. Yet these datasets are time-consuming and labor …

Tresnet: High performance gpu-dedicated architecture

T Ridnik, H Lawen, A Noy… - proceedings of the …, 2021 - openaccess.thecvf.com
Many deep learning models, developed in recent years, reach higher ImageNet accuracy
than ResNet50, with fewer or comparable FLOPs count. While FLOPs are often seen as a …

Mixed autoencoder for self-supervised visual representation learning

K Chen, Z Liu, L Hong, H Xu, Z Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
Masked Autoencoder (MAE) has demonstrated superior performance on various vision tasks
via randomly masking image patches and reconstruction. However, effective data …

How far pre-trained models are from neural collapse on the target dataset informs their transferability

Z Wang, Y Luo, L Zheng, Z Huang… - Proceedings of the …, 2023 - openaccess.thecvf.com
This paper focuses on model transferability estimation, ie, assessing the performance of pre-
trained models on a downstream task without performing fine-tuning. Motivated by the …

Beyond synthetic noise: Deep learning on controlled noisy labels

L Jiang, D Huang, M Liu… - … conference on machine …, 2020 - proceedings.mlr.press
Performing controlled experiments on noisy data is essential in understanding deep
learning across noise levels. Due to the lack of suitable datasets, previous research has only …