Efficient Navigation Among Movable Obstacles using a Mobile Manipulator via Hierarchical Policy Learning
By: Taegeun Yang , Jiwoo Hwang , Jeil Jeong and more
Potential Business Impact:
Robot learns to push obstacles out of the way.
We propose a hierarchical reinforcement learning (HRL) framework for efficient Navigation Among Movable Obstacles (NAMO) using a mobile manipulator. Our approach combines interaction-based obstacle property estimation with structured pushing strategies, facilitating the dynamic manipulation of unforeseen obstacles while adhering to a pre-planned global path. The high-level policy generates pushing commands that consider environmental constraints and path-tracking objectives, while the low-level policy precisely and stably executes these commands through coordinated whole-body movements. Comprehensive simulation-based experiments demonstrate improvements in performing NAMO tasks, including higher success rates, shortened traversed path length, and reduced goal-reaching times, compared to baselines. Additionally, ablation studies assess the efficacy of each component, while a qualitative analysis further validates the accuracy and reliability of the real-time obstacle property estimation.
Similar Papers
NAMO-LLM: Efficient Navigation Among Movable Obstacles with Large Language Model Guidance
Robotics
Helps robots move through crowded spaces.
Hierarchical Learning-Enhanced MPC for Safe Crowd Navigation with Heterogeneous Constraints
Robotics
Robots navigate tricky, changing places better.
NAMOUnc: Navigation Among Movable Obstacles with Decision Making on Uncertainty Interval
Robotics
Helps robots safely move around moving things.