Hypo: Hyperspherical out-of-distribution generalization

Y Ming, H Bai, J Katz-Samuels, Y Li - arXiv preprint arXiv:2402.07785, 2024 - arxiv.org
Out-of-distribution (OOD) generalization is critical for machine learning models deployed in
the real world. However, achieving this can be fundamentally challenging, as it requires the …

Cifar-10-warehouse: Broad and more realistic testbeds in model generalization analysis

X Sun, X Leng, Z Wang, Y Yang, Z Huang… - arXiv preprint arXiv …, 2023 - arxiv.org
Analyzing model performance in various unseen environments is a critical research problem
in the machine learning community. To study this problem, it is important to construct a …

Matrix information theory for self-supervised learning

Y Zhang, Z Tan, J Yang, W Huang, Y Yuan - arXiv preprint arXiv …, 2023 - arxiv.org
The maximum entropy encoding framework provides a unified perspective for many non-
contrastive learning methods like SimSiam, Barlow Twins, and MEC. Inspired by this …

Aha: Human-assisted out-of-distribution generalization and detection

H Bai, J Zhang, R Nowak - arXiv preprint arXiv:2410.08000, 2024 - arxiv.org
Modern machine learning models deployed often encounter distribution shifts in real-world
applications, manifesting as covariate or semantic out-of-distribution (OOD) shifts. These …

Rethinking the Representation in Federated Unsupervised Learning with Non-IID Data

X Liao, W Liu, C Chen, P Zhou, F Yu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Federated learning achieves effective performance in modeling decentralized data. In
practice client data are not well-labeled which makes it potential for federated unsupervised …

FroSSL: Frobenius Norm Minimization for Efficient Multiview Self-supervised Learning

O Skean, A Dhakal, N Jacobs… - European Conference on …, 2025 - Springer
Self-supervised learning (SSL) is a popular paradigm for representation learning. Recent
multiview methods can be classified as sample-contrastive, dimension-contrastive, or …

Improving forward compatibility in class incremental learning by increasing representation rank and feature richness

J Kim, W Lee, M Eo, W Rhee - Neural Networks, 2025 - Elsevier
Abstract Class Incremental Learning (CIL) constitutes a pivotal subfield within continual
learning, aimed at enabling models to progressively learn new classification tasks while …

Label smoothing regularization-based no hyperparameter domain generalization

Y Wang, X Wu, XY Liu, F Chu, H Liu, Z Han - Knowledge-Based Systems, 2024 - Elsevier
Abstract Domain generalization learns from one or multiple source domains. It aims to
extract a domain-invariant model that can be employed in an unknown target domain …

Kernel Language Entropy: Fine-grained Uncertainty Quantification for LLMs from Semantic Similarities

A Nikitin, J Kossen, Y Gal, P Marttinen - arXiv preprint arXiv:2405.20003, 2024 - arxiv.org
Uncertainty quantification in Large Language Models (LLMs) is crucial for applications
where safety and reliability are important. In particular, uncertainty can be used to improve …

Learning to Embed Distributions via Maximum Kernel Entropy

O Kachaiev, S Recanatesi - arXiv preprint arXiv:2408.00549, 2024 - arxiv.org
Empirical data can often be considered as samples from a set of probability distributions.
Kernel methods have emerged as a natural approach for learning to classify these …