Collision avoidance from monocular vision trained with novel view synthesis
By: Valentin Tordjman--Levavasseur, Stéphane Caron
Potential Business Impact:
Robot sees obstacles, avoids crashing into them.
Collision avoidance can be checked in explicit environment models such as elevation maps or occupancy grids, yet integrating such models with a locomotion policy requires accurate state estimation. In this work, we consider the question of collision avoidance from an implicit environment model. We use monocular RGB images as inputs and train a collisionavoidance policy from photorealistic images generated by 2D Gaussian splatting. We evaluate the resulting pipeline in realworld experiments under velocity commands that bring the robot on an intercept course with obstacles. Our results suggest that RGB images can be enough to make collision-avoidance decisions, both in the room where training data was collected and in out-of-distribution environments.
Similar Papers
MonoMPC: Monocular Vision Based Navigation with Learned Collision Model and Risk-Aware Model Predictive Control
Robotics
Robot navigates safely through messy places.
RGB-Event Fusion with Self-Attention for Collision Prediction
Robotics
Helps robots avoid crashing into things.
Computer vision training dataset generation for robotic environments using Gaussian splatting
CV and Pattern Recognition
Creates realistic fake pictures for robots to learn.