Physically-based Lighting Generation for Robotic Manipulation
By: Shutong Jin , Lezhong Wang , Ben Temming and more
Potential Business Impact:
Makes robots learn new tasks in different lights.
In this paper, we propose the first framework that leverages physically-based inverse rendering for novel lighting generation on existing real-world human demonstrations of robotic manipulation tasks. Specifically, inverse rendering decomposes the first frame in each demonstration into geometric (surface normal, depth) and material (albedo, roughness, metallic) properties, which are then used to render appearance changes under different lighting sources. To improve efficiency and maintain consistency across each generated sequence, we fine-tune Stable Video Diffusion on robot execution videos for temporal lighting propagation. We evaluate our framework by measuring the visual quality of the generated sequences, assessing its effectiveness in improving the imitation learning policy performance (38.75\%) under six unseen real-world lighting conditions, and conduct ablation studies on individual modules of the proposed framework. We further showcase three downstream applications enabled by the proposed framework: background generation, object texture generation and distractor positioning. The code for the framework will be made publicly available.
Similar Papers
Physically-based Lighting Augmentation for Robotic Manipulation
Robotics
Teaches robots to see in any light.
Physically Controllable Relighting of Photographs
Graphics
Changes how lights look in real pictures.
Image Generation as a Visual Planner for Robotic Manipulation
CV and Pattern Recognition
Lets computers plan robot actions by watching videos.