A Graph-Based Reinforcement Learning Approach with Frontier Potential Based Reward for Safe Cluttered Environment Exploration
By: Gabriele Calzolari , Vidya Sumathy , Christoforos Kanellakis and more
Potential Business Impact:
Robots explore tricky places without crashing.
Autonomous exploration of cluttered environments requires efficient exploration strategies that guarantee safety against potential collisions with unknown random obstacles. This paper presents a novel approach combining a graph neural network-based exploration greedy policy with a safety shield to ensure safe navigation goal selection. The network is trained using reinforcement learning and the proximal policy optimization algorithm to maximize exploration efficiency while reducing the safety shield interventions. However, if the policy selects an infeasible action, the safety shield intervenes to choose the best feasible alternative, ensuring system consistency. Moreover, this paper proposes a reward function that includes a potential field based on the agent's proximity to unexplored regions and the expected information gain from reaching them. Overall, the approach investigated in this paper merges the benefits of the adaptability of reinforcement learning-driven exploration policies and the guarantee ensured by explicit safety mechanisms. Extensive evaluations in simulated environments demonstrate that the approach enables efficient and safe exploration in cluttered environments.
Similar Papers
Platform-Agnostic Reinforcement Learning Framework for Safe Exploration of Cluttered Environments with Graph Attention
Robotics
Helps robots explore dangerous places safely.
Bridging Deep Reinforcement Learning and Motion Planning for Model-Free Navigation in Cluttered Environments
Robotics
Helps robots explore tricky places better.
Probabilistic Shielding for Safe Reinforcement Learning
Machine Learning (Stat)
Keeps robots safe while they learn new tasks.