Advances in variational inference

C Zhang, J Bütepage, H Kjellström… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
Many modern unsupervised or semi-supervised machine learning algorithms rely on
Bayesian probabilistic models. These models are usually intractable and thus require …

Variational inference: A review for statisticians

DM Blei, A Kucukelbir, JD McAuliffe - Journal of the American …, 2017 - Taylor & Francis
One of the core problems of modern statistics is to approximate difficult-to-compute
probability densities. This problem is especially important in Bayesian statistics, which …

Monte carlo gradient estimation in machine learning

S Mohamed, M Rosca, M Figurnov, A Mnih - Journal of Machine Learning …, 2020 - jmlr.org
This paper is a broad and accessible survey of the methods we have at our disposal for
Monte Carlo gradient estimation in machine learning and across the statistical sciences: the …

Virtual adversarial training: a regularization method for supervised and semi-supervised learning

T Miyato, S Maeda, M Koyama… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
We propose a new regularization method based on virtual adversarial loss: a new measure
of local smoothness of the conditional label distribution given input. Virtual adversarial loss …

The concrete distribution: A continuous relaxation of discrete random variables

CJ Maddison, A Mnih, YW Teh - arXiv preprint arXiv:1611.00712, 2016 - arxiv.org
The reparameterization trick enables optimizing large scale stochastic computation graphs
via gradient descent. The essence of the trick is to refactor each stochastic node into a …

Rebar: Low-variance, unbiased gradient estimates for discrete latent variable models

G Tucker, A Mnih, CJ Maddison… - Advances in …, 2017 - proceedings.neurips.cc
Learning in models with discrete latent variables is challenging due to high variance
gradient estimators. Generally, approaches have relied on control variates to reduce the …

Variational inference for monte carlo objectives

A Mnih, D Rezende - International Conference on Machine …, 2016 - proceedings.mlr.press
Recent progress in deep latent variable models has largely been driven by the development
of flexible and scalable variational inference methods. Variational training of this type …

The generalized reparameterization gradient

FR Ruiz, TRC AUEB, D Blei - Advances in neural …, 2016 - proceedings.neurips.cc
The reparameterization gradient has become a widely used method to obtain Monte Carlo
gradients to optimize the variational objective. However, this technique does not easily apply …

Gradient estimation with stochastic softmax tricks

M Paulus, D Choi, D Tarlow… - Advances in neural …, 2020 - proceedings.neurips.cc
Abstract The Gumbel-Max trick is the basis of many relaxed gradient estimators. These
estimators are easy to implement and low variance, but the goal of scaling them …

Provable convergence guarantees for black-box variational inference

J Domke, R Gower, G Garrigos - Advances in neural …, 2024 - proceedings.neurips.cc
Black-box variational inference is widely used in situations where there is no proof that its
stochastic optimization succeeds. We suggest this is due to a theoretical gap in existing …