Model Identification Adaptive Control with $ρ$-POMDP Planning
By: Michelle Ho, Arec Jamgochian, Mykel J. Kochenderfer
Potential Business Impact:
Teaches robots to learn and control things better.
Accurate system modeling is crucial for safe, effective control, as misidentification can lead to accumulated errors, especially under partial observability. We address this problem by formulating informative input design and model identification adaptive control (MIAC) as belief space planning problems, modeled as partially observable Markov decision processes with belief-dependent rewards ($\rho$-POMDPs). We treat system parameters as hidden state variables that must be localized while simultaneously controlling the system. We solve this problem with an adapted belief-space iterative Linear Quadratic Regulator (BiLQR). We demonstrate it on fully and partially observable tasks for cart-pole and steady aircraft flight domains. Our method outperforms baselines such as regression, filtering, and local optimal control methods, even under instantaneous disturbances to system parameters.
Similar Papers
Learning global control of underactuated systems with Model-Based Reinforcement Learning
Robotics
Teaches robots to learn new tasks faster.
Estimating unknown dynamics and cost as a bilinear system with Koopman-based Inverse Optimal Control
Systems and Control
Helps robots learn to move perfectly.
Fast RLS Identification Leveraging the Linearized System Sparsity: Predictive Cost Adaptive Control for Quadrotors
Systems and Control
Drones fly better by learning how they move.