Optimistic mirror descent in saddle-point problems: Going the extra (gradient) mile

P Mertikopoulos, B Lecouat, H Zenati, CS Foo… - arXiv preprint arXiv …, 2018 - arxiv.org
Owing to their connection with generative adversarial networks (GANs), saddle-point
problems have recently attracted considerable interest in machine learning and beyond. By …

A tail-index analysis of stochastic gradient noise in deep neural networks

U Simsekli, L Sagun… - … Conference on Machine …, 2019 - proceedings.mlr.press
The gradient noise (GN) in the stochastic gradient descent (SGD) algorithm is often
considered to be Gaussian in the large data regime by assuming that the classical central …

Adaptive learning in continuous games: Optimal regret bounds and convergence to nash equilibrium

YG Hsieh, K Antonakopoulos… - … on Learning Theory, 2021 - proceedings.mlr.press
In game-theoretic learning, several agents are simultaneously following their individual
interests, so the environment is non-stationary from each player's perspective. In this context …

On gradient-based learning in continuous games

E Mazumdar, LJ Ratliff, SS Sastry - SIAM Journal on Mathematics of Data …, 2020 - SIAM
We introduce a general framework for competitive gradient-based learning that
encompasses a wide breadth of multiagent learning algorithms, and analyze the limiting …

Stochastic heavy ball

S Gadat, F Panloup, S Saadane - 2018 - projecteuclid.org
This paper deals with a natural stochastic optimization procedure derived from the so-called
Heavy-ball method differential equation, which was introduced by Polyak in the 1960s with …

Mirrored langevin dynamics

YP Hsieh, A Kavis, P Rolland… - Advances in Neural …, 2018 - proceedings.neurips.cc
We consider the problem of sampling from constrained distributions, which has posed
significant challenges to both non-asymptotic analysis and algorithmic design. We propose …

Distributed asynchronous optimization with unbounded delays: How slow can you go?

Z Zhou, P Mertikopoulos, N Bambos… - International …, 2018 - proceedings.mlr.press
One of the most widely used optimization methods for large-scale machine learning
problems is distributed asynchronous stochastic gradient descent (DASGD). However, a key …

A unified stochastic approximation framework for learning in games

P Mertikopoulos, YP Hsieh, V Cevher - Mathematical Programming, 2024 - Springer
We develop a flexible stochastic approximation framework for analyzing the long-run
behavior of learning in games (both continuous and finite). The proposed analysis template …

An adaptive mirror-prox method for variational inequalities with singular operators

K Antonakopoulos, V Belmega… - Advances in Neural …, 2019 - proceedings.neurips.cc
Lipschitz continuity is a central requirement for achieving the optimal O (1/T) rate of
convergence in monotone, deterministic variational inequalities (a setting that includes …

Riemannian game dynamics

P Mertikopoulos, WH Sandholm - Journal of Economic Theory, 2018 - Elsevier
We study a class of evolutionary game dynamics defined by balancing a gain determined by
the game's payoffs against a cost of motion that captures the difficulty with which the …