Statistical learning theory for control: A finite-sample perspective

A Tsiamis, I Ziemann, N Matni… - IEEE Control Systems …, 2023 - ieeexplore.ieee.org
Learning algorithms have become an integral component to modern engineering solutions.
Examples range from self-driving cars and recommender systems to finance and even …

Neural operators for bypassing gain and control computations in PDE backstep**

L Bhan, Y Shi, M Krstic - IEEE Transactions on Automatic …, 2023 - ieeexplore.ieee.org
We introduce a framework for eliminating the computation of controller gain functions in
partial differential equation (PDE) control. We learn the nonlinear operator from the plant …

[HTML][HTML] Neural operators of backstep** controller and observer gain functions for reaction–diffusion PDEs

M Krstic, L Bhan, Y Shi - Automatica, 2024 - Elsevier
Unlike ODEs, whose models involve system matrices and whose controllers involve vector
or matrix gains, PDE models involve functions in those roles—functional coefficients …

Data-enabled policy optimization for direct adaptive learning of the LQR

F Zhao, F Dörfler, A Chiuso, K You - arxiv preprint arxiv:2401.14871, 2024 - arxiv.org
Direct data-driven design methods for the linear quadratic regulator (LQR) mainly use offline
or episodic data batches, and their online adaptation has been acknowledged as an open …

On the optimization landscape of dynamic output feedback linear quadratic control

J Duan, W Cao, Y Zheng, L Zhao - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
The convergence of policy gradient algorithms hinges on the optimization landscape of the
underlying optimal control problem. Theoretical insights into these algorithms can often be …

Global convergence of policy gradient primal–dual methods for risk-constrained LQRs

F Zhao, K You, T Başar - IEEE Transactions on Automatic …, 2023 - ieeexplore.ieee.org
While the techniques in optimal control theory are often model-based, the policy optimization
(PO) approach directly optimizes the performance metric of interest. Even though it has been …

Fast trac: A parameter-free optimizer for lifelong reinforcement learning

A Muppidi, Z Zhang, H Yang - Advances in Neural …, 2025 - proceedings.neurips.cc
A key challenge in lifelong reinforcement learning (RL) is the loss of plasticity, where
previous learning progress hinders an agent's adaptation to new tasks. While regularization …

Model-free learning with heterogeneous dynamical systems: A federated LQR approach

H Wang, LF Toso, A Mitra, J Anderson - arxiv preprint arxiv:2308.11743, 2023 - arxiv.org
We study a model-free federated linear quadratic regulator (LQR) problem where M agents
with unknown, distinct yet similar dynamics collaboratively learn an optimal policy to …

Revisiting LQR control from the perspective of receding-horizon policy gradient

X Zhang, T Başar - IEEE Control Systems Letters, 2023 - ieeexplore.ieee.org
We revisit in this letter the discrete-time linear quadratic regulator (LQR) problem from the
perspective of receding-horizon policy gradient (RHPG), a newly developed model-free …

A reinforcement learning look at risk-sensitive linear quadratic gaussian control

L Cui, T Basar, ZP Jiang - Learning for Dynamics and Control …, 2023 - proceedings.mlr.press
In this paper, we propose a robust reinforcement learning method for a class of linear
discrete-time systems to handle model mismatches that may be induced by sim-to-real gap …