Learning the Riccati solution operator for time-varying LQR via Deep Operator Networks
Jun Chen, Umberto Biccari, Junmin Wang
TLDR
A DeepONet framework learns the Riccati solution operator for time-varying LQR, offering fast, accurate, and stable optimal control.
Key contributions
- Proposes a DeepONet-based operator to replace repeated numerical solutions of Riccati equations in LQR.
- Enables fast online evaluation of optimal feedbacks by shifting computation to a one-time offline learning stage.
- Provides theoretical guarantees on error propagation and preserves closed-loop system stability under approximation.
- Demonstrates high accuracy, strong generalization, and significant speedups for time-varying LQR problems.
Why it matters
This work offers a significant advancement for real-time and parametric optimal control by replacing computationally intensive Riccati equation solving with a learned, fast-evaluating operator. It provides crucial theoretical guarantees, ensuring reliability and stability for data-driven control systems.
Original Abstract
We propose a computational framework for replacing the repeated numerical solution of differential Riccati equations in finite-horizon Linear Quadratic Regulator (LQR) problems by a learned operator surrogate. Instead of solving a nonlinear matrix-valued differential equation for each new system instance, we construct offline an approximation of the associated solution operator mapping time-dependent system parameters to the Riccati trajectory. The resulting model enables fast online evaluation of approximate optimal feedbacks across a wide class of systems, thereby shifting the computational burden from repeated numerical integration to a one-time learning stage. From a theoretical perspective, we establish control-theoretic guarantees for this operator-based approximation. In particular, we derive bounds quantifying how operator approximation errors propagate to feedback performance, trajectory accuracy, and cost suboptimality, and we prove that exponential stability of the closed-loop system is preserved under sufficiently accurate operator approximation. These results provide a framework to assess the reliability of data-driven approximations in optimal control. On the computational side, we design tailored DeepONet architectures for matrix-valued, time-dependent problems and introduce a progressive learning strategy to address scalability with respect to the system dimension. Numerical experiments on both time-invariant and time-varying LQR problems demonstrate that the proposed approach achieves high accuracy and strong generalization across a wide range of system configurations, while delivering substantial computational speedups compared to classical solvers. The method offers an effective and scalable alternative for parametric and real-time optimal control applications.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.