Goal Reaching with Eikonal-Constrained Hierarchical Quasimetric Reinforcement Learning
By: Vittorio Giammarino, Ahmed H. Qureshi
Goal-Conditioned Reinforcement Learning (GCRL) mitigates the difficulty of reward design by framing tasks as goal reaching rather than maximizing hand-crafted reward signals. In this setting, the optimal goal-conditioned value function naturally forms a quasimetric, motivating Quasimetric RL (QRL), which constrains value learning to quasimetric mappings and enforces local consistency through discrete, trajectory-based constraints. We propose Eikonal-Constrained Quasimetric RL (Eik-QRL), a continuous-time reformulation of QRL based on the Eikonal Partial Differential Equation (PDE). This PDE-based structure makes Eik-QRL trajectory-free, requiring only sampled states and goals, while improving out-of-distribution generalization. We provide theoretical guarantees for Eik-QRL and identify limitations that arise under complex dynamics. To address these challenges, we introduce Eik-Hierarchical QRL (Eik-HiQRL), which integrates Eik-QRL into a hierarchical decomposition. Empirically, Eik-HiQRL achieves state-of-the-art performance in offline goal-conditioned navigation and yields consistent gains over QRL in manipulation tasks, matching temporal-difference methods.
Similar Papers
Physics-informed Value Learner for Offline Goal-Conditioned Reinforcement Learning
Machine Learning (CS)
Teaches robots to navigate complex places safely.
Multistep Quasimetric Learning for Scalable Goal-conditioned Reinforcement Learning
Machine Learning (CS)
Helps robots learn long tasks from watching.
Multistep Quasimetric Learning for Scalable Goal-conditioned Reinforcement Learning
Machine Learning (CS)
Teaches robots to do tasks by watching.