Sensorized gripper for human demonstrations
By: Sri Harsha Turlapati , Gautami Golani , Mohammad Zaidi Ariffin and more
Potential Business Impact:
Robots learn to build things by watching humans.
Ease of programming is a key factor in making robots ubiquitous in unstructured environments. In this work, we present a sensorized gripper built with off-the-shelf parts, used to record human demonstrations of a box in box assembly task. With very few trials of short interval timings each, we show that a robot can repeat the task successfully. We adopt a Cartesian approach to robot motion generation by computing the joint space solution while concurrently solving for the optimal robot position, to maximise manipulability. The statistics of the human demonstration are extracted using Gaussian Mixture Models (GMM) and the robot is commanded using impedance control.
Similar Papers
A High-Force Gripper with Embedded Multimodal Sensing for Powerful and Perception Driven Grasping
Robotics
Robot hands can now grip much heavier things.
Imitation Learning with Precisely Labeled Human Demonstrations
Robotics
Teaches robots to learn from human actions.
Optimizing Robot Programming: Mixed Reality Gripper Control
Human-Computer Interaction
Teaches robots new jobs faster with a special controller.