Agreement-on-the-line: Predicting the performance of neural networks under distribution shift

C Baek, Y Jiang, A Raghunathan… - Advances in Neural …, 2022 - proceedings.neurips.cc
Recently, Miller et al. showed that a model's in-distribution (ID) accuracy has a strong linear
correlation with its out-of-distribution (OOD) accuracy, on several OOD benchmarks, a …

Predicting with confidence on unseen distributions

D Guillory, V Shankar, S Ebrahimi… - Proceedings of the …, 2021 - openaccess.thecvf.com
Recent work has shown that the accuracy of machine learning models can vary substantially
when evaluated on a distribution that even slightly differs from that of the training data. As a …

Leveraging unlabeled data to predict out-of-distribution performance

S Garg, S Balakrishnan, ZC Lipton… - arXiv preprint arXiv …, 2022 - arxiv.org
Real-world machine learning deployments are characterized by mismatches between the
source (training) and target (test) distributions that may cause performance drops. In this …

Probvlm: Probabilistic adapter for frozen vison-language models

U Upadhyay, S Karthik, M Mancini… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large-scale vision-language models (VLMs) like CLIP successfully find correspondences
between images and text. Through the standard deterministic mapping process, an image or …

Detecting errors and estimating accuracy on unlabeled data with self-training ensembles

J Chen, F Liu, B Avci, X Wu… - Advances in Neural …, 2021 - proceedings.neurips.cc
When a deep learning model is deployed in the wild, it can encounter test data drawn from
distributions different from the training data distribution and suffer drop in performance. For …

Gnnevaluator: Evaluating gnn performance on unseen graphs without labels

X Zheng, M Zhang, C Chen, S Molaei… - Advances in Neural …, 2024 - proceedings.neurips.cc
Evaluating the performance of graph neural networks (GNNs) is an essential task for
practical GNN model deployment and serving, as deployed GNNs face significant …

A survey on evaluation of out-of-distribution generalization

H Yu, J Liu, X Zhang, J Wu, P Cui - arXiv preprint arXiv:2403.01874, 2024 - arxiv.org
Machine learning models, while progressively advanced, rely heavily on the IID assumption,
which is often unfulfilled in practice due to inevitable distribution shifts. This renders them …

Predicting out-of-distribution error with the projection norm

Y Yu, Z Yang, A Wei, Y Ma… - … Conference on Machine …, 2022 - proceedings.mlr.press
We propose a metric—Projection Norm—to predict a model's performance on out-of-
distribution (OOD) data without access to ground truth labels. Projection Norm first uses …

Came: Contrastive automated model evaluation

R Peng, Q Duan, H Wang, J Ma… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract The Automated Model Evaluation (AutoEval) framework entertains the possibility of
evaluating a trained machine learning model without resorting to a labeled testing set …

Characterizing out-of-distribution error via optimal transport

Y Lu, Y Qin, R Zhai, A Shen, K Chen… - Advances in …, 2023 - proceedings.neurips.cc
Abstract Out-of-distribution (OOD) data poses serious challenges in deployed machine
learning models, so methods of predicting a model's performance on OOD data without …