Score: 0

Data-Driven Inverse Optimal Control for Continuous-Time Nonlinear Systems

Published: March 12, 2025 | arXiv ID: 2503.09090v2

By: Hamed Jabbari Asl, Eiji Uchibe

Potential Business Impact:

Helps robots learn how to do tasks by watching.

Business Areas:
Industrial Automation Manufacturing, Science and Engineering

This paper introduces a novel model-free and a partially model-free algorithm for inverse optimal control (IOC), also known as inverse reinforcement learning (IRL), aimed at estimating the cost function of continuous-time nonlinear deterministic systems. Using the input-state trajectories of an expert agent, the proposed algorithms separately utilize control policy information and the Hamilton-Jacobi-Bellman equation to estimate different sets of cost function parameters. This approach allows the algorithms to achieve broader applicability while maintaining a model-free framework. Also, the model-free algorithm reduces complexity compared to existing methods, as it requires solving a forward optimal control problem only once during initialization. Furthermore, in our partially model-free algorithm, this step can be bypassed entirely for systems with known input dynamics. Simulation results demonstrate the effectiveness and efficiency of our algorithms, highlighting their potential for real-world deployment in autonomous systems and robotics.

Page Count
10 pages

Category
Electrical Engineering and Systems Science:
Systems and Control