Ratatouille: Imitation Learning Ingredients for Real-world Social Robot Navigation
By: James R. Han , Mithun Vanniasinghe , Hshmat Sahak and more
Potential Business Impact:
Robots learn to walk safely around people.
Scaling Reinforcement Learning to in-the-wild social robot navigation is both data-intensive and unsafe, since policies must learn through direct interaction and inevitably encounter collisions. Offline Imitation learning (IL) avoids these risks by collecting expert demonstrations safely, training entirely offline, and deploying policies zero-shot. However, we find that naively applying Behaviour Cloning (BC) to social navigation is insufficient; achieving strong performance requires careful architectural and training choices. We present Ratatouille, a pipeline and model architecture that, without changing the data, reduces collisions per meter by 6 times and improves success rate by 3 times compared to naive BC. We validate our approach in both simulation and the real world, where we collected over 11 hours of data on a dense university campus. We further demonstrate qualitative results in a public food court. Our findings highlight that thoughtful IL design, rather than additional data, can substantially improve safety and reliability in real-world social navigation. Video: https://youtu.be/tOdLTXsaYLQ. Code will be released after acceptance.
Similar Papers
R2BC: Multi-Agent Imitation Learning from Single-Agent Demonstrations
Robotics
Teaches robot teams by showing one robot at a time.
From Imitation to Optimization: A Comparative Study of Offline Learning for Autonomous Driving
Machine Learning (CS)
Teaches self-driving cars to avoid crashes.
Counterfactual Behavior Cloning: Offline Imitation Learning from Imperfect Human Demonstrations
Robotics
Robots learn better from human mistakes.