An overview of multi-agent reinforcement learning from game theoretical perspective

Y Yang, J Wang - arXiv preprint arXiv:2011.00583, 2020 - arxiv.org
Following the remarkable success of the AlphaGO series, 2019 was a booming year that
witnessed significant advances in multi-agent reinforcement learning (MARL) techniques …

[图书][B] Random dynamical systems

L Arnold, CKRT Jones, K Mischaikow, G Raugel… - 1995 - Springer
The theory of random dynamical systems continues, extends, and unites various
developments in probability theory and dynamical systems. Roughly speaking, a random …

Bome! bilevel optimization made easy: A simple first-order approach

B Liu, M Ye, S Wright, P Stone… - Advances in neural …, 2022 - proceedings.neurips.cc
Bilevel optimization (BO) is useful for solving a variety of important machine learning
problems including but not limited to hyperparameter optimization, meta-learning, continual …

[PDF][PDF] Gradient descent only converges to minimizers

JD Lee, M Simchowitz, MI Jordan… - Conference on learning …, 2016 - proceedings.mlr.press
Gradient Descent Only Converges to Minimizers Page 1 JMLR: Workshop and Conference
Proceedings vol 49:1–12, 2016 Gradient Descent Only Converges to Minimizers Jason D. Lee …

The limit points of (optimistic) gradient descent in min-max optimization

C Daskalakis, I Panageas - Advances in neural information …, 2018 - proceedings.neurips.cc
Motivated by applications in Optimization, Game Theory, and the training of Generative
Adversarial Networks, the convergence properties of first order methods in min-max …

First-order methods almost always avoid strict saddle points

JD Lee, I Panageas, G Piliouras, M Simchowitz… - Mathematical …, 2019 - Springer
We establish that first-order methods avoid strict saddle points for almost all initializations.
Our results apply to a wide variety of first-order methods, including (manifold) gradient …

Understanding the unstable convergence of gradient descent

K Ahn, J Zhang, S Sra - International Conference on …, 2022 - proceedings.mlr.press
Most existing analyses of (stochastic) gradient descent rely on the condition that for $ L $-
smooth costs, the step size is less than $2/L $. However, many works have observed that in …

Implicit learning dynamics in stackelberg games: Equilibria characterization, convergence analysis, and empirical study

T Fiez, B Chasnov, L Ratliff - International Conference on …, 2020 - proceedings.mlr.press
Contemporary work on learning in continuous games has commonly overlooked the
hierarchical decision-making structure present in machine learning problems formulated as …

Texts in Applied Mathematics 2

JE Marsden, L Sirovich, SS Antman, G Iooss, P Holmes… - 2003 - Springer
In this book we will study equations of the following form x= f (x, t; µ),(0.0. 1) and x↦→ g (x;
µ),(0.0. 2) with x∈ U⊂ Rn, t∈ R1, and µ∈ V⊂ Rp where U and V are open sets in Rn and …

[图书][B] Algebra

S Lang - 2012 - books.google.com
" Lang's Algebra changed the way graduate algebra is taught, retaining classical topics but
introducing language and ways of thinking from category theory and homological algebra. It …