Change is hard: A closer look at subpopulation shift

Y Yang, H Zhang, D Katabi, M Ghassemi - arXiv preprint arXiv:2302.12254, 2023 - arxiv.org
Machine learning models often perform poorly on subgroups that are underrepresented in
the training data. Yet, little is understood on the variation in mechanisms that cause …

Breeds: Benchmarks for subpopulation shift

S Santurkar, D Tsipras, A Madry - arXiv preprint arXiv:2008.04859, 2020 - arxiv.org
We develop a methodology for assessing the robustness of models to subpopulation shift---
specifically, their ability to generalize to novel data subpopulations that were not observed …

Umix: Improving importance weighting for subpopulation shift via uncertainty-aware mixup

Z Han, Z Liang, F Yang, L Liu, L Li… - Advances in …, 2022 - proceedings.neurips.cc
Subpopulation shift widely exists in many real-world machine learning applications, referring
to the training and test distributions containing the same subpopulation groups but varying in …

A fine-grained analysis on distribution shift

O Wiles, S Gowal, F Stimberg, S Alvise-Rebuffi… - arXiv preprint arXiv …, 2021 - arxiv.org
Robustness to distribution shifts is critical for deploying machine learning models in the real
world. Despite this necessity, there has been little work in defining the underlying …

Can you rely on your model evaluation? improving model evaluation with synthetic test data

B van Breugel, N Seedat, F Imrie… - Advances in Neural …, 2024 - proceedings.neurips.cc
Evaluating the performance of machine learning models on diverse and underrepresented
subgroups is essential for ensuring fairness and reliability in real-world applications …

Examining and combating spurious features under distribution shift

C Zhou, X Ma, P Michel… - … Conference on Machine …, 2021 - proceedings.mlr.press
A central goal of machine learning is to learn robust representations that capture the
fundamental relationship between inputs and output labels. However, minimizing training …

Metashift: A dataset of datasets for evaluating contextual distribution shifts and training conflicts

W Liang, J Zou - arXiv preprint arXiv:2202.06523, 2022 - arxiv.org
Understanding the performance of machine learning models across diverse data
distributions is critically important for reliable applications. Motivated by this, there is a …

Model patching: Closing the subgroup performance gap with data augmentation

K Goel, A Gu, Y Li, C Ré - arXiv preprint arXiv:2008.06775, 2020 - arxiv.org
Classifiers in machine learning are often brittle when deployed. Particularly concerning are
models with inconsistent performance on specific subgroups of a class, eg, exhibiting …

Bias mimicking: A simple sampling approach for bias mitigation

M Qraitem, K Saenko… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Prior work has shown that Visual Recognition datasets frequently underrepresent bias
groups B (eg Female) within class labels Y (eg Programmers). This dataset bias can lead to …

Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization

S Sagawa, PW Koh, TB Hashimoto, P Liang - arXiv preprint arXiv …, 2019 - arxiv.org
Overparameterized neural networks can be highly accurate on average on an iid test set yet
consistently fail on atypical groups of the data (eg, by learning spurious correlations that …