[HTML][HTML] To compress or not to compress—self-supervised learning and information theory: A review

R Shwartz Ziv, Y LeCun - Entropy, 2024 - mdpi.com
Deep neural networks excel in supervised learning tasks but are constrained by the need for
extensive labeled data. Self-supervised learning emerges as a promising alternative …

When does label smoothing help?

R Müller, S Kornblith, GE Hinton - Advances in neural …, 2019 - proceedings.neurips.cc
The generalization and learning speed of a multi-class neural network can often be
significantly improved by using soft targets that are a weighted average of the hard targets …

Opening the black box of deep neural networks via information

R Shwartz-Ziv, N Tishby - arXiv preprint arXiv:1703.00810, 2017 - arxiv.org
Despite their great success, there is still no comprehensive theoretical understanding of
learning with Deep Neural Networks (DNNs) or their inner organization. Previous work …

On information plane analyses of neural network classifiers—A review

BC Geiger - IEEE Transactions on Neural Networks and …, 2021 - ieeexplore.ieee.org
We review the current literature concerned with information plane (IP) analyses of neural
network (NN) classifiers. While the underlying information bottleneck theory and the claim …

How does information bottleneck help deep learning?

K Kawaguchi, Z Deng, X Ji… - … Conference on Machine …, 2023 - proceedings.mlr.press
Numerous deep learning algorithms have been inspired by and understood via the notion of
information bottleneck, where unnecessary information is (often implicitly) minimized while …

Hrel: Filter pruning based on high relevance between activation maps and class labels

CH Sarvani, M Ghorai, SR Dubey, SHS Basha - Neural Networks, 2022 - Elsevier
This paper proposes an Information Bottleneck theory based filter pruning method that uses
a statistical measure called Mutual Information (MI). The MI between filters and class labels …

Where is the information in a deep neural network?

A Achille, G Paolini, S Soatto - arXiv preprint arXiv:1905.12213, 2019 - arxiv.org
Whatever information a deep neural network has gleaned from training data is encoded in
its weights. How this information affects the response of the network to future data remains …

Understanding convolutional neural networks with information theory: An initial exploration

S Yu, K Wickstrøm, R Jenssen… - IEEE transactions on …, 2020 - ieeexplore.ieee.org
A novel functional estimator for Rényi's α-entropy and its multivariate extension was recently
proposed in terms of the normalized eigenspectrum of a Hermitian matrix of the projected …

Information flow in deep neural networks

R Shwartz-Ziv - arXiv preprint arXiv:2202.06749, 2022 - arxiv.org
Although deep neural networks have been immensely successful, there is no
comprehensive theoretical understanding of how they work or are structured. As a result …

[PDF][PDF] Information-Theoretic Methods in Deep Neural Networks: Recent Advances and Emerging Opportunities.

S Yu, LGS Giraldo, JC Príncipe - IJCAI, 2021 - ijcai.org
We present a review on the recent advances and emerging opportunities around the theme
of analyzing deep neural networks (DNNs) with information-theoretic methods. We first …