Estimating high order gradients of the data distribution by denoising

C Meng, Y Song, W Li, S Ermon - Advances in Neural …, 2021 - proceedings.neurips.cc
The first order derivative of a data density can be estimated efficiently by denoising score
matching, and has become an important component in many applications, such as image …

Stochastic runge-kutta accelerates langevin monte carlo and beyond

X Li, Y Wu, L Mackey… - Advances in neural …, 2019 - proceedings.neurips.cc
Abstract Sampling with Markov chain Monte Carlo methods typically amounts to discretizing
some continuous-time dynamics with numerical integration. In this paper, we establish the …

On stochastic gradient langevin dynamics with dependent data streams: The fully nonconvex case

NH Chau, É Moulines, M Rásonyi, S Sabanis… - SIAM Journal on …, 2021 - SIAM
We consider the problem of sampling from a target distribution, which is not necessarily log-
concave, in the context of empirical risk minimization and stochastic optimization as …

Nonasymptotic analysis of Stochastic Gradient Hamiltonian Monte Carlo under local conditions for nonconvex optimization

OD Akyildiz, S Sabanis - Journal of Machine Learning Research, 2024 - jmlr.org
We provide a nonasymptotic analysis of the convergence of the stochastic gradient
Hamiltonian Monte Carlo (SGHMC) to a target measure in Wasserstein-2 distance without …

Nonasymptotic estimates for stochastic gradient Langevin dynamics under local conditions in nonconvex optimization

Y Zhang, ÖD Akyildiz, T Damoulas… - Applied Mathematics & …, 2023 - Springer
In this paper, we are concerned with a non-asymptotic analysis of sampling algorithms used
in nonconvex optimization. In particular, we obtain non-asymptotic estimates in Wasserstein …

On the ergodicity, bias and asymptotic normality of randomized midpoint sampling method

Y He, K Balasubramanian… - Advances in Neural …, 2020 - proceedings.neurips.cc
The randomized midpoint method, proposed by (Shen and Lee, 2019), has emerged as an
optimal discretization procedure for simulating the continuous time underdamped Langevin …

Taming neural networks with tusla: Nonconvex learning via adaptive stochastic gradient langevin algorithms

A Lovas, I Lytras, M Rásonyi, S Sabanis - SIAM Journal on Mathematics of …, 2023 - SIAM
Artificial neural networks (ANNs) are typically highly nonlinear systems which are finely
tuned via the optimization of their associated, nonconvex loss functions. In many cases, the …

On the posterior distribution in denoising: Application to uncertainty quantification

H Manor, T Michaeli - arXiv preprint arXiv:2309.13598, 2023 - arxiv.org
Denoisers play a central role in many applications, from noise suppression in low-grade
imaging sensors, to empowering score-based generative models. The latter category of …

Optimized population monte carlo

V Elvira, E Chouzenoux - IEEE Transactions on Signal …, 2022 - ieeexplore.ieee.org
Adaptive importance sampling (AIS) methods are increasingly used for the approximation of
distributions and related intractable integrals in the context of Bayesian inference …

[HTML][HTML] Kinetic Langevin MCMC Sampling Without Gradient Lipschitz Continuity-the Strongly Convex Case

T Johnston, I Lytras, S Sabanis - Journal of Complexity, 2024 - Elsevier
In this article we consider sampling from log concave distributions in Hamiltonian setting,
without assuming that the objective gradient is globally Lipschitz. We propose two …