Bimanual 3D Hand Motion and Articulation Forecasting in Everyday Images
By: Aditya Prakash, David Forsyth, Saurabh Gupta
Potential Business Impact:
Lets computers guess how hands will move.
We tackle the problem of forecasting bimanual 3D hand motion & articulation from a single image in everyday settings. To address the lack of 3D hand annotations in diverse settings, we design an annotation pipeline consisting of a diffusion model to lift 2D hand keypoint sequences to 4D hand motion. For the forecasting model, we adopt a diffusion loss to account for the multimodality in hand motion distribution. Extensive experiments across 6 datasets show the benefits of training on diverse data with imputed labels (14% improvement) and effectiveness of our lifting (42% better) & forecasting (16.4% gain) models, over the best baselines, especially in zero-shot generalization to everyday images.
Similar Papers
Diffusion-based 3D Hand Motion Recovery with Intuitive Physics
CV and Pattern Recognition
Makes computer hands move realistically when touching things.
The Invisible EgoHand: 3D Hand Forecasting through EgoBody Pose Estimation
CV and Pattern Recognition
Predicts where hands will move, even when hidden.
Novel Diffusion Models for Multimodal 3D Hand Trajectory Prediction
CV and Pattern Recognition
Predicts hand movements using many senses.