HumanoidGen: Data Generation for Bimanual Dexterous Manipulation via LLM Reasoning
By: Zhi Jing , Siyuan Yang , Jicong Ao and more
Potential Business Impact:
Teaches human-like robots to do tasks.
For robotic manipulation, existing robotics datasets and simulation benchmarks predominantly cater to robot-arm platforms. However, for humanoid robots equipped with dual arms and dexterous hands, simulation tasks and high-quality demonstrations are notably lacking. Bimanual dexterous manipulation is inherently more complex, as it requires coordinated arm movements and hand operations, making autonomous data collection challenging. This paper presents HumanoidGen, an automated task creation and demonstration collection framework that leverages atomic dexterous operations and LLM reasoning to generate relational constraints. Specifically, we provide spatial annotations for both assets and dexterous hands based on the atomic operations, and perform an LLM planner to generate a chain of actionable spatial constraints for arm movements based on object affordances and scenes. To further improve planning ability, we employ a variant of Monte Carlo tree search to enhance LLM reasoning for long-horizon tasks and insufficient annotation. In experiments, we create a novel benchmark with augmented scenarios to evaluate the quality of the collected data. The results show that the performance of the 2D and 3D diffusion policies can scale with the generated dataset. Project page is https://openhumanoidgen.github.io.
Similar Papers
GenDexHand: Generative Simulation for Dexterous Hands
Robotics
Creates robot hands that can do many tasks.
GENMANIP: LLM-driven Simulation for Generalizable Instruction-Following Manipulation
Robotics
Makes robots learn new tasks from simple instructions.
DexMan: Learning Bimanual Dexterous Manipulation from Human and Generated Videos
Robotics
Robots learn to do tasks by watching videos.