Stochastic variance reduction for variational inequality methods

A Alacaoglu, Y Malitsky - Conference on Learning Theory, 2022 - proceedings.mlr.press
We propose stochastic variance reduced algorithms for solving convex-concave saddle
point problems, monotone variational inequalities, and monotone inclusions. Our framework …

On solving minimax optimization locally: A follow-the-ridge approach

Y Wang, G Zhang, J Ba - arXiv preprint arXiv:1910.07512, 2019 - arxiv.org
Many tasks in modern machine learning can be formulated as finding equilibria in\emph
{sequential} games. In particular, two-player zero-sum sequential games, also known as …

Logan: Latent optimisation for generative adversarial networks

Y Wu, J Donahue, D Balduzzi, K Simonyan… - arXiv preprint arXiv …, 2019 - arxiv.org
Training generative adversarial networks requires balancing of delicate adversarial
dynamics. Even with careful tuning, training may diverge or end up in a bad equilibrium with …

Competitive gradient descent

F Schäfer, A Anandkumar - Advances in Neural Information …, 2019 - proceedings.neurips.cc
We introduce a new algorithm for the numerical computation of Nash equilibria of
competitive two-player games. Our method is a natural generalization of gradient descent to …

Stable opponent shaping in differentiable games

A Letcher, J Foerster, D Balduzzi, T Rocktäschel… - arXiv preprint arXiv …, 2018 - arxiv.org
A growing number of learning methods are actually differentiable games whose players
optimise multiple, interdependent objectives in parallel--from GANs and intrinsic curiosity to …

Differentiable game mechanics

A Letcher, D Balduzzi, S Racaniere, J Martens… - Journal of Machine …, 2019 - jmlr.org
Deep learning is built on the foundational guarantee that gradient descent on an objective
function converges to local minima. Unfortunately, this guarantee fails in settings, such as …

Competitive physics informed networks

Q Zeng, Y Kothari, SH Bryngelson, F Schäfer - arXiv preprint arXiv …, 2022 - arxiv.org
Neural networks can be trained to solve partial differential equations (PDEs) by using the
PDE residual as the loss function. This strategy is called" physics-informed neural …

Finding mixed nash equilibria of generative adversarial networks

YP Hsieh, C Liu, V Cevher - International Conference on …, 2019 - proceedings.mlr.press
Generative adversarial networks (GANs) are known to achieve the state-of-the-art
performance on various generative tasks, but these results come at the expense of a …

Training generative adversarial networks by solving ordinary differential equations

C Qin, Y Wu, JT Springenberg… - Advances in …, 2020 - proceedings.neurips.cc
Abstract The instability of Generative Adversarial Network (GAN) training has frequently
been attributed to gradient descent. Consequently, recent methods have aimed to tailor the …

Policy gradient methods find the nash equilibrium in n-player general-sum linear-quadratic games

B Hambly, R Xu, H Yang - Journal of Machine Learning Research, 2023 - jmlr.org
We consider a general-sum N-player linear-quadratic game with stochastic dynamics over a
finite horizon and prove the global convergence of the natural policy gradient method to the …