User-friendly introduction to PAC-Bayes bounds

P Alquier - Foundations and Trends® in Machine Learning, 2024 - nowpublishers.com
Aggregated predictors are obtained by making a set of basic predictors vote according to
some weights, that is, to some probability distribution. Randomized predictors are obtained …

How does information bottleneck help deep learning?

K Kawaguchi, Z Deng, X Ji… - … Conference on Machine …, 2023 - proceedings.mlr.press
Numerous deep learning algorithms have been inspired by and understood via the notion of
information bottleneck, where unnecessary information is (often implicitly) minimized while …

Reasoning about generalization via conditional mutual information

T Steinke, L Zakynthinou - Conference on Learning Theory, 2020 - proceedings.mlr.press
We provide an information-theoretic framework for studying the generalization properties of
machine learning algorithms. Our framework ties together existing approaches, including …

Tightening mutual information-based bounds on generalization error

Y Bu, S Zou, VV Veeravalli - IEEE Journal on Selected Areas in …, 2020 - ieeexplore.ieee.org
An information-theoretic upper bound on the generalization error of supervised learning
algorithms is derived. The bound is constructed in terms of the mutual information between …

Information-theoretic generalization bounds for SGLD via data-dependent estimates

J Negrea, M Haghifam, GK Dziugaite… - Advances in …, 2019 - proceedings.neurips.cc
In this work, we improve upon the stepwise analysis of noisy iterative learning algorithms
initiated by Pensia, Jog, and Loh (2018) and recently extended by Bu, Zou, and Veeravalli …

When is memorization of irrelevant training data necessary for high-accuracy learning?

G Brown, M Bun, V Feldman, A Smith… - Proceedings of the 53rd …, 2021 - dl.acm.org
Modern machine learning models are complex and frequently encode surprising amounts of
information about individual inputs. In extreme cases, complex models appear to memorize …

Sharpened generalization bounds based on conditional mutual information and an application to noisy, iterative algorithms

M Haghifam, J Negrea, A Khisti… - Advances in …, 2020 - proceedings.neurips.cc
The information-theoretic framework of Russo and Zou (2016) and Xu and Raginsky (2017)
provides bounds on the generalization error of a learning algorithm in terms of the mutual …

Predictive information accelerates learning in rl

KH Lee, I Fischer, A Liu, Y Guo, H Lee… - Advances in …, 2020 - proceedings.neurips.cc
Abstract The Predictive Information is the mutual information between the past and the
future, I (Xpast; Xfuture). We hypothesize that capturing the predictive information is useful in …

Information-theoretic generalization bounds for black-box learning algorithms

H Harutyunyan, M Raginsky… - Advances in Neural …, 2021 - proceedings.neurips.cc
We derive information-theoretic generalization bounds for supervised learning algorithms
based on the information contained in predictions rather than in the output of the training …

Chaining mutual information and tightening generalization bounds

A Asadi, E Abbe, S Verdú - Advances in Neural Information …, 2018 - proceedings.neurips.cc
Bounding the generalization error of learning algorithms has a long history, which yet falls
short in explaining various generalization successes including those of deep learning. Two …