[图书][B] Control systems and reinforcement learning

S Meyn - 2022 - books.google.com
A high school student can create deep Q-learning code to control her robot, without any
understanding of the meaning of'deep'or'Q', or why the code sometimes fails. This book is …

Distributed online optimization in dynamic environments using mirror descent

S Shahrampour, A Jadbabaie - IEEE Transactions on Automatic …, 2017 - ieeexplore.ieee.org
This work addresses decentralized online optimization in nonstationary environments. A
network of agents aim to track the minimizer of a global, time-varying, and convex function …

Accelerated mirror descent in continuous and discrete time

W Krichene, A Bayen… - Advances in neural …, 2015 - proceedings.neurips.cc
We study accelerated mirror descent dynamics in continuous and discrete time. Combining
the original continuous-time motivation of mirror descent with a recent ODE interpretation of …

Stochastic modified equations and dynamics of stochastic gradient algorithms i: Mathematical foundations

Q Li, C Tai, E Weinan - Journal of Machine Learning Research, 2019 - jmlr.org
We develop the mathematical foundations of the stochastic modified equations (SME)
framework for analyzing the dynamics of stochastic gradient algorithms, where the latter is …

A dual approach for optimal algorithms in distributed optimization over networks

CA Uribe, S Lee, A Gasnikov… - 2020 Information theory …, 2020 - ieeexplore.ieee.org
We study dual-based algorithms for distributed convex optimization problems over networks,
where the objective is to minimize a sum Σ i= 1 mfi (z) of functions over in a network. We …

Direct Runge-Kutta discretization achieves acceleration

J Zhang, A Mokhtari, S Sra… - Advances in neural …, 2018 - proceedings.neurips.cc
We study gradient-based optimization methods obtained by directly discretizing a second-
order ordinary differential equation (ODE) related to the continuous limit of Nesterov's …

Mirrored langevin dynamics

YP Hsieh, A Kavis, P Rolland… - Advances in Neural …, 2018 - proceedings.neurips.cc
We consider the problem of sampling from constrained distributions, which has posed
significant challenges to both non-asymptotic analysis and algorithmic design. We propose …

Reparameterizing mirror descent as gradient descent

E Amid, MKK Warmuth - Advances in Neural Information …, 2020 - proceedings.neurips.cc
Most of the recent successful applications of neural networks have been based on training
with gradient descent updates. However, for some small networks, other mirror descent …

On the convergence of gradient-like flows with noisy gradient input

P Mertikopoulos, M Staudigl - SIAM Journal on Optimization, 2018 - SIAM
In view of solving convex optimization problems with noisy gradient input, we analyze the
asymptotic behavior of gradient-like flows under stochastic disturbances. Specifically, we …

Distributed randomized gradient-free mirror descent algorithm for constrained optimization

Z Yu, DWC Ho, D Yuan - IEEE Transactions on Automatic …, 2021 - ieeexplore.ieee.org
This article is concerned with the multiagent optimization problem. A distributed randomized
gradient-free mirror descent (DRGFMD) method is developed by introducing a randomized …