A review of Shannon and differential entropy rate estimation

A Feutrill, M Roughan - Entropy, 2021 - mdpi.com
In this paper, we present a review of Shannon and differential entropy rate estimation
techniques. Entropy rate, which measures the average information gain from a stochastic …

Empirical estimation of information measures: A literature guide

S Verdú - Entropy, 2019 - mdpi.com
We give a brief survey of the literature on the empirical estimation of entropy, differential
entropy, relative entropy, mutual information and related information measures. While those …

Polynomial methods in statistical inference: theory and practice

Y Wu, P Yang - Foundations and Trends® in …, 2020 - nowpublishers.com
This survey provides an exposition of a suite of techniques based on the theory of
polynomials, collectively referred to as polynomial methods, which have recently been …

Entropy–and Distance-Regularized Attention Improves Low-Resource Neural Machine Translation

A Araabi, V Niculae, C Monz - … of the 16th Conference of the …, 2024 - aclanthology.org
Transformer-based models in Neural Machine Translation (NMT) rely heavily on multi-head
attention for capturing dependencies within and across source and target sequences. In …

Optimal prediction of markov chains with and without spectral gap

Y Han, S Jana, Y Wu - Advances in Neural Information …, 2021 - proceedings.neurips.cc
We study the following learning problem with dependent data: Given a trajectory of length $
n $ from a stationary Markov chain with $ k $ states, the goal is to predict the distribution of …

Quantifying and Estimating the Predictability Upper Bound of Univariate Numeric Time Series

J Mohammed, MH Böhlen, S Helmer - Proceedings of the 30th ACM …, 2024 - dl.acm.org
The intrinsic predictability of a given time series indicates how well an (ideal) algorithm
could potentially predict it when trained on the time series data. Being able to compute the …

Cross entropy of neural language models at infinity—a new bound of the entropy rate

S Takahashi, K Tanaka-Ishii - Entropy, 2018 - mdpi.com
Neural language models have drawn a lot of attention for their strong ability to predict
natural language text. In this paper, we estimate the entropy rate of natural language with …

Estimating the fundamental limits is easier than achieving the fundamental limits

J Jiao, Y Han, I Fischer-Hwang… - IEEE Transactions on …, 2019 - ieeexplore.ieee.org
We show through case studies that it is easier to estimate the fundamental limits of data
processing than to construct the explicit algorithms to achieve those limits. Focusing on …

Minimax redundancy for Markov chains with large state space

K Tatwawadi, J Jiao, T Weissman - 2018 IEEE International …, 2018 - ieeexplore.ieee.org
For any Markov source, there exist universal codes whose normalized codelength
approaches the Shannon limit asymptotically as the number of samples goes to infinity. This …

Optimal prediction of Markov chains with and without spectral gap

Y Han, S Jana, Y Wu - IEEE Transactions on Information …, 2023 - ieeexplore.ieee.org
We study the following learning problem with dependent data: Observing a trajectory of
length from a stationary Markov chain with states, the goal is to predict the next state. For …