Data-Driven Inverse Optimal Control for Continuous-Time Nonlinear Systems
By: Hamed Jabbari Asl, Eiji Uchibe
Potential Business Impact:
Helps robots learn how to do tasks by watching.
This paper introduces a novel model-free and a partially model-free algorithm for inverse optimal control (IOC), also known as inverse reinforcement learning (IRL), aimed at estimating the cost function of continuous-time nonlinear deterministic systems. Using the input-state trajectories of an expert agent, the proposed algorithms separately utilize control policy information and the Hamilton-Jacobi-Bellman equation to estimate different sets of cost function parameters. This approach allows the algorithms to achieve broader applicability while maintaining a model-free framework. Also, the model-free algorithm reduces complexity compared to existing methods, as it requires solving a forward optimal control problem only once during initialization. Furthermore, in our partially model-free algorithm, this step can be bypassed entirely for systems with known input dynamics. Simulation results demonstrate the effectiveness and efficiency of our algorithms, highlighting their potential for real-world deployment in autonomous systems and robotics.
Similar Papers
Estimating unknown dynamics and cost as a bilinear system with Koopman-based Inverse Optimal Control
Systems and Control
Helps robots learn to move perfectly.
A Model-Free Data-Driven Algorithm for Continuous-Time Control
Optimization and Control
Teaches robots to control things without knowing how they work.
Inverse Reinforcement Learning via Convex Optimization
Machine Learning (CS)
Teaches computers to learn what people want.