On the System Theoretic Offline Learning of Continuous-Time LQR with Exogenous Disturbances
By: Sayak Mukherjee, Ramij R. Hossain, Mahantesh Halappanavar
Potential Business Impact:
Helps robots learn to control things even with surprises.
We analyze offline designs of linear quadratic regulator (LQR) strategies with uncertain disturbances. First, we consider the scenario where the exogenous variable can be estimated in a controlled environment, and subsequently, consider a more practical and challenging scenario where it is unknown in a stochastic setting. Our approach builds on the fundamental learning-based framework of adaptive dynamic programming (ADP), combined with a Lyapunov-based analytical methodology to design the algorithms and derive sample-based approximations motivated from the Markov decision process (MDP)-based approaches. For the scenario involving non-measurable disturbances, we further establish stability and convergence guarantees for the learned control gains under sample-based approximations. The overall methodology emphasizes simplicity while providing rigorous guarantees. Finally, numerical experiments focus on the intricacies and validations for the design of offline continuous-time LQR with exogenous disturbances.
Similar Papers
Optimal Output Feedback Learning Control for Discrete-Time Linear Quadratic Regulation
Systems and Control
Teaches robots to learn how to control things.
Direct Data Driven Control Using Noisy Measurements
Systems and Control
Learns to control machines from messy data.
A model free approach for continuous-time optimal tracking control with unknown user-define cost and constrained control input via advantage function
Systems and Control
Teaches robots to control things without knowing how.