Stochastic optimization with heavy-tailed noise via accelerated gradient clipping

E Gorbunov, M Danilova… - Advances in Neural …, 2020 - proceedings.neurips.cc
In this paper, we propose a new accelerated stochastic first-order method called clipped-
SSTM for smooth convex stochastic optimization with heavy-tailed distributed noise in …

Recent theoretical advances in non-convex optimization

M Danilova, P Dvurechensky, A Gasnikov… - … and Probability: With a …, 2022 - Springer
Motivated by recent increased interest in optimization algorithms for non-convex
optimization in application to training deep neural networks and other optimization problems …

The power of first-order smooth optimization for black-box non-smooth problems

A Gasnikov, A Novitskii, V Novitskii… - arXiv preprint arXiv …, 2022 - arxiv.org
Gradient-free/zeroth-order methods for black-box convex optimization have been
extensively studied in the last decade with the main focus on oracle calls complexity. In this …

Optimal decentralized distributed algorithms for stochastic convex optimization

E Gorbunov, D Dvinskikh, A Gasnikov - arXiv preprint arXiv:1911.07363, 2019 - arxiv.org
We consider stochastic convex optimization problems with affine constraints and develop
several methods using either primal or dual approach to solve it. In the primal case, we use …

Decentralized local stochastic extra-gradient for variational inequalities

A Beznosikov, P Dvurechenskii… - Advances in …, 2022 - proceedings.neurips.cc
We consider distributed stochastic variational inequalities (VIs) on unbounded domains with
the problem data that is heterogeneous (non-IID) and distributed across many devices. We …

Decentralized and parallel primal and dual accelerated methods for stochastic convex programming problems

D Dvinskikh, A Gasnikov - Journal of Inverse and Ill-posed Problems, 2021 - degruyter.com
We introduce primal and dual stochastic gradient oracle methods for decentralized convex
optimization problems. Both for primal and dual oracles, the proposed methods are optimal …

Recent theoretical advances in decentralized distributed convex optimization

E Gorbunov, A Rogozin, A Beznosikov… - … and Probability: With a …, 2022 - Springer
In the last few years, the theory of decentralized distributed convex optimization has made
significant progress. The lower bounds on communications rounds and oracle calls have …

Randomized gradient-free methods in convex optimization

A Gasnikov, D Dvinskikh, P Dvurechensky… - Encyclopedia of …, 2023 - Springer
Consider a convex optimization problem min x∈ Q⊆ Rd f (x)(1) with convex feasible set Q
and convex objective f possessing the zeroth-order (gradient/derivativefree) oracle [83]. The …

The “Black-Box” Optimization Problem: Zero-Order Accelerated Stochastic Method via Kernel Approximation

A Lobanov, N Bashirov, A Gasnikov - Journal of Optimization Theory and …, 2024 - Springer
In this paper, we study the standard formulation of an optimization problem when the
computation of gradient is not available. Such a problem can be classified as a “black box” …

Stochastic three points method for unconstrained smooth minimization

EH Bergou, E Gorbunov, P Richtarik - SIAM Journal on Optimization, 2020 - SIAM
In this paper we consider the unconstrained minimization problem of a smooth function in
R^n in a setting where only function evaluations are possible. We design a novel …