Fine-tuning can distort pretrained features and underperform out-of-distribution

A Kumar, A Raghunathan, R Jones, T Ma… - arXiv preprint arXiv …, 2022 - arxiv.org
When transferring a pretrained model to a downstream task, two popular methods are full
fine-tuning (updating all the model parameters) and linear probing (updating only the last …

On the opportunities and risks of foundation models

R Bommasani, DA Hudson, E Adeli, R Altman… - arXiv preprint arXiv …, 2021 - arxiv.org
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …

Provable guarantees for self-supervised deep learning with spectral contrastive loss

JZ HaoChen, C Wei, A Gaidon… - Advances in Neural …, 2021 - proceedings.neurips.cc
Recent works in self-supervised learning have advanced the state-of-the-art by relying on
the contrastive learning paradigm, which learns representations by pushing positive pairs, or …

Connect, not collapse: Explaining contrastive learning for unsupervised domain adaptation

K Shen, RM Jones, A Kumar, SM Xie… - International …, 2022 - proceedings.mlr.press
We consider unsupervised domain adaptation (UDA), where labeled data from a source
domain (eg, photos) and unlabeled data from a target domain (eg, sketches) are used to …

Divide and contrast: Source-free domain adaptation via adaptive contrastive learning

Z Zhang, W Chen, H Cheng, Z Li… - Advances in Neural …, 2022 - proceedings.neurips.cc
We investigate a practical domain adaptation task, called source-free domain adaptation
(SFUDA), where the source pretrained model is adapted to the target domain without access …

Change is hard: A closer look at subpopulation shift

Y Yang, H Zhang, D Katabi, M Ghassemi - arXiv preprint arXiv:2302.12254, 2023 - arxiv.org
Machine learning models often perform poorly on subgroups that are underrepresented in
the training data. Yet, little is understood on the variation in mechanisms that cause …

Complementary benefits of contrastive learning and self-training under distribution shift

S Garg, A Setlur, Z Lipton… - Advances in …, 2024 - proceedings.neurips.cc
Self-training and contrastive learning have emerged as leading techniques for incorporating
unlabeled data, both under distribution shift (unsupervised domain adaptation) and when it …

Invariant and transportable representations for anti-causal domain shifts

Y Jiang, V Veitch - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Real-world classification problems must contend with domain shift, the (potential) mismatch
between the domain where a model is deployed and the domain (s) where the training data …

Beyond separability: Analyzing the linear transferability of contrastive representations to related subpopulations

JZ HaoChen, C Wei, A Kumar… - Advances in neural …, 2022 - proceedings.neurips.cc
Contrastive learning is a highly effective method for learning representations from unlabeled
data. Recent works show that contrastive representations can transfer across domains …

Transferring fairness under distribution shifts via fair consistency regularization

B An, Z Che, M Ding, F Huang - Advances in Neural …, 2022 - proceedings.neurips.cc
The increasing reliance on ML models in high-stakes tasks has raised a major concern
about fairness violations. Although there has been a surge of work that improves algorithmic …