Demonstration Sidetracks: Categorizing Systematic Non-Optimality in Human Demonstrations
By: Shijie Fang , Hang Yu , Qidi Fang and more
Potential Business Impact:
Teaches robots better from human mistakes.
Learning from Demonstration (LfD) is a popular approach for robots to acquire new skills, but most LfD methods suffer from imperfections in human demonstrations. Prior work typically treats these suboptimalities as random noise. In this paper we study non-optimal behaviors in non-expert demonstrations and show that they are systematic, forming what we call demonstration sidetracks. Using a public space study with 40 participants performing a long-horizon robot task, we recreated the setup in simulation and annotated all demonstrations. We identify four types of sidetracks (Exploration, Mistake, Alignment, Pause) and one control pattern (one-dimension control). Sidetracks appear frequently across participants, and their temporal and spatial distribution is tied to task context. We also find that users' control patterns depend on the control interface. These insights point to the need for better models of suboptimal demonstrations to improve LfD algorithms and bridge the gap between lab training and real-world deployment. All demonstrations, infrastructure, and annotations are available at https://github.com/AABL-Lab/Human-Demonstration-Sidetracks.
Similar Papers
Active Robot Curriculum Learning from Online Human Demonstrations
Robotics
Teaches robots better by asking for help smartly.
Training People to Reward Robots
Robotics
Teaches robots to learn better from humans.
An Alignment-Based Approach to Learning Motions from Demonstrations
Robotics
Teaches robots new moves by watching humans.