Vision-based Perception System for Automated Delivery Robot-Pedestrians Interactions
By: Ergi Tushe, Bilal Farooq
Potential Business Impact:
Helps robots safely navigate crowded sidewalks.
The integration of Automated Delivery Robots (ADRs) into pedestrian-heavy urban spaces introduces unique challenges in terms of safe, efficient, and socially acceptable navigation. We develop the complete pipeline for a single vision sensor based multi-pedestrian detection and tracking, pose estimation, and monocular depth perception. Leveraging the real-world MOT17 dataset sequences, this study demonstrates how integrating human-pose estimation and depth cues enhances pedestrian trajectory prediction and identity maintenance, even under occlusions and dense crowds. Results show measurable improvements, including up to a 10% increase in identity preservation (IDF1), a 7% improvement in multiobject tracking accuracy (MOTA), and consistently high detection precision exceeding 85%, even in challenging scenarios. Notably, the system identifies vulnerable pedestrian groups supporting more socially aware and inclusive robot behaviour.
Similar Papers
Vision-Aided Online A* Path Planning for Efficient and Safe Navigation of Service Robots
Robotics
Robot sees important things, not just obstacles.
A Multi-Drone Multi-View Dataset and Deep Learning Framework for Pedestrian Detection and Tracking
CV and Pattern Recognition
Tracks people from many moving cameras.
Realistic pedestrian-driver interaction modelling using multi-agent RL with human perceptual-motor constraints
Artificial Intelligence
Makes self-driving cars understand people better.