Last-iterate convergence in extensive-form games

CW Lee, C Kroer, H Luo - Advances in Neural Information …, 2021 - proceedings.neurips.cc
Regret-based algorithms are highly efficient at finding approximate Nash equilibria in
sequential games such as poker games. However, most regret-based algorithms, including …

Faster game solving via predictive blackwell approachability: Connecting regret matching and mirror descent

G Farina, C Kroer, T Sandholm - … of the AAAI Conference on Artificial …, 2021 - ojs.aaai.org
Blackwell approachability is a framework for reasoning about repeated games with vector-
valued payoffs. We introduce predictive Blackwell approachability, where an estimate of the …

Block-coordinate methods and restarting for solving extensive-form games

D Chakrabarti, J Diakonikolas… - Advances in Neural …, 2023 - proceedings.neurips.cc
Coordinate descent methods are popular in machine learning and optimization for their
simple sparse updates and excellent practical performance. In the context of large-scale …

Meta-learning in games

K Harris, I Anagnostides, G Farina, M Khodak… - arXiv preprint arXiv …, 2022 - arxiv.org
In the literature on game-theoretic equilibrium finding, focus has mainly been on solving a
single game in isolation. In practice, however, strategic interactions--ranging from routing …

Last-iterate convergence rates for min-max optimization: Convergence of hamiltonian gradient descent and consensus optimization

J Abernethy, KA Lai, A Wibisono - Algorithmic Learning …, 2021 - proceedings.mlr.press
While classic work in convex-concave min-max optimization relies on average-iterate
convergence results, the emergence of nonconvex applications such as training Generative …

Optimistic regret minimization for extensive-form games via dilated distance-generating functions

G Farina, C Kroer, T Sandholm - Advances in neural …, 2019 - proceedings.neurips.cc
We study the performance of optimistic regret-minimization algorithms for both minimizing
regret in, and computing Nash equilibria of, zero-sum extensive-form games. In order to …

Last-iterate convergence rates for min-max optimization

J Abernethy, KA Lai, A Wibisono - arXiv preprint arXiv:1906.02027, 2019 - arxiv.org
While classic work in convex-concave min-max optimization relies on average-iterate
convergence results, the emergence of nonconvex applications such as training Generative …

First-order methods for Wasserstein distributionally robust MDP

JG Clement, C Kroer - International Conference on Machine …, 2021 - proceedings.mlr.press
Markov decision processes (MDPs) are known to be sensitive to parameter specification.
Distributionally robust MDPs alleviate this issue by allowing for\textit {ambiguity sets} which …

Scalable first-order methods for robust mdps

J Grand-Clément, C Kroer - Proceedings of the AAAI Conference on …, 2021 - ojs.aaai.org
Abstract Robust Markov Decision Processes (MDPs) are a powerful framework for modeling
sequential decision making problems with model uncertainty. This paper proposes the first …

Solving optimization problems with blackwell approachability

J Grand-Clément, C Kroer - Mathematics of Operations …, 2024 - pubsonline.informs.org
In this paper, we propose a new algorithm for solving convex-concave saddle-point
problems using regret minimization in the repeated game framework. To do so, we introduce …