Imitation Learning Based on Disentangled Representation Learning of Behavioral Characteristics
By: Ryoga Oishi, Sho Sakaino, Toshiaki Tsuji
Potential Business Impact:
Robots change how they move based on your words.
In the field of robot learning, coordinating robot actions through language instructions is becoming increasingly feasible. However, adapting actions to human instructions remains challenging, as such instructions are often qualitative and require exploring behaviors that satisfy varying conditions. This paper proposes a motion generation model that adapts robot actions in response to modifier directives human instructions imposing behavioral conditions during task execution. The proposed method learns a mapping from modifier directives to actions by segmenting demonstrations into short sequences, assigning weakly supervised labels corresponding to specific modifier types. We evaluated our method in wiping and pick and place tasks. Results show that it can adjust motions online in response to modifier directives, unlike conventional batch-based methods that cannot adapt during execution.
Similar Papers
Deep Sensorimotor Control by Imitating Predictive Models of Human Motion
Robotics
Robots learn to move by watching humans.
From Generated Human Videos to Physically Plausible Robot Trajectories
Robotics
Robots copy human moves from fake videos.
Structured Imitation Learning of Interactive Policies through Inverse Games
Robotics
Teaches robots to work with people.