Toward a theoretical foundation of policy optimization for learning control policies

B Hu, K Zhang, N Li, M Mesbahi… - Annual Review of …, 2023 - annualreviews.org
Gradient-based methods have been widely used for system design and optimization in
diverse application domains. Recently, there has been a renewed interest in studying …

Derivative-free optimization methods

J Larson, M Menickelly, SM Wild - Acta Numerica, 2019 - cambridge.org
In many optimization problems arising from scientific, engineering and artificial intelligence
applications, objective and constraint functions are available only as the output of a black …

Gradient-free methods for deterministic and stochastic nonsmooth nonconvex optimization

T Lin, Z Zheng, M Jordan - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Nonsmooth nonconvex optimization problems broadly emerge in machine learning and
business decision making, whereas two core challenges impede the development of …

Global planning for contact-rich manipulation via local smoothing of quasi-dynamic contact models

T Pang, HJT Suh, L Yang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
The empirical success of reinforcement learning (RL) in contact-rich manipulation leaves
much to be understood from a model-based perspective, where the key difficulties are often …

Complexity of finding stationary points of nonconvex nonsmooth functions

J Zhang, H Lin, S Jegelka, S Sra… - … on Machine Learning, 2020 - proceedings.mlr.press
We provide the first non-asymptotic analysis for finding stationary points of nonsmooth,
nonconvex functions. In particular, we study the class of Hadamard semi-differentiable …

Deterministic nonsmooth nonconvex optimization

M Jordan, G Kornowski, T Lin… - The Thirty Sixth …, 2023 - proceedings.mlr.press
We study the complexity of optimizing nonsmooth nonconvex Lipschitz functions by
producing $(\delta,\epsilon) $-Goldstein stationary points. Several recent works have …

Bundled gradients through contact via randomized smoothing

HJT Suh, T Pang, R Tedrake - IEEE Robotics and Automation …, 2022 - ieeexplore.ieee.org
The empirical success of derivative-free methods in reinforcement learning for planning
through contact seems at odds with the perceived fragility of classical gradient-based …

A gradient sampling method with complexity guarantees for Lipschitz functions in high and low dimensions

D Davis, D Drusvyatskiy, YT Lee… - Advances in neural …, 2022 - proceedings.neurips.cc
Abstract Zhang et al.(ICML 2020) introduced a novel modification of Goldstein's classical
subgradient method, with an efficiency guarantee of $ O (\varepsilon^{-4}) $ for minimizing …

Complexity of Derivative-Free Policy Optimization for Structured Control

X Guo, D Keivan, G Dullerud… - Advances in Neural …, 2024 - proceedings.neurips.cc
The applications of direct policy search in reinforcement learning and continuous control
have received increasing attention. In this work, we present novel theoretical results on the …

On the finite-time complexity and practical computation of approximate stationarity concepts of lipschitz functions

L Tian, K Zhou, AMC So - International Conference on …, 2022 - proceedings.mlr.press
We report a practical finite-time algorithmic scheme to compute approximately stationary
points for nonconvex nonsmooth Lipschitz functions. In particular, we are interested in two …