SensHRPS: Sensing Comfortable Human-Robot Proxemics and Personal Space With Eye-Tracking
By: Nadezhda Kushina , Ko Watanabe , Aarthi Kannan and more
Potential Business Impact:
Robot learns how close to stand to make you comfy.
Social robots must adjust to human proxemic norms to ensure user comfort and engagement. While prior research demonstrates that eye-tracking features reliably estimate comfort in human-human interactions, their applicability to interactions with humanoid robots remains unexplored. In this study, we investigate user comfort with the robot "Ameca" across four experimentally controlled distances (0.5 m to 2.0 m) using mobile eye-tracking and subjective reporting (N=19). We evaluate multiple machine learning and deep learning models to estimate comfort based on gaze features. Contrary to previous human-human studies where Transformer models excelled, a Decision Tree classifier achieved the highest performance (F1-score = 0.73), with minimum pupil diameter identified as the most critical predictor. These findings suggest that physiological comfort thresholds in human-robot interaction differ from human-human dynamics and can be effectively modeled using interpretable logic.
Similar Papers
Using Physiological Measures, Gaze, and Facial Expressions to Model Human Trust in a Robot Partner
Robotics
Helps robots know when people trust them.
Beyond the Plane: A 3D Representation of Human Personal Space for Socially-Aware Robotics
Robotics
Robots learn to respect your personal bubble in 3D.
Examining the legibility of humanoid robot arm movements in a pointing task
Robotics
Helps robots show where they're going.