作者
Jingliang Duan, Wenhan Cao, Yang Zheng, Lin Zhao
发表日期
2023/5/12
期刊
IEEE Transactions on Automatic Control
卷号
69
期号
2
页码范围
920-935
出版商
IEEE
简介
The convergence of policy gradient algorithms hinges on the optimization landscape of the underlying optimal control problem. Theoretical insights into these algorithms can often be acquired from analyzing those of linear quadratic control. However, most of the existing literature only considers the optimization landscape for static full-state or output feedback policies (controllers). In this article, we investigate the more challenging case of dynamic output-feedback policies for linear quadratic regulation (abbreviated as dLQR ), which is prevalent in practice but has a rather complicated optimization landscape. We first show how the dLQR cost varies with the coordinate transformation of the dynamic controller, and then, derive the optimal transformation for a given observable stabilizing controller. One of our core results is the uniqueness of the stationary point of dLQR when it is observable, which provides an …
引用总数
学术搜索中的文章