Score: 0

Physically-based Lighting Generation for Robotic Manipulation

Published: August 2, 2025 | arXiv ID: 2508.01442v2

By: Shutong Jin , Lezhong Wang , Ben Temming and more

Potential Business Impact:

Makes robots learn new tasks in different lights.

In this paper, we propose the first framework that leverages physically-based inverse rendering for novel lighting generation on existing real-world human demonstrations of robotic manipulation tasks. Specifically, inverse rendering decomposes the first frame in each demonstration into geometric (surface normal, depth) and material (albedo, roughness, metallic) properties, which are then used to render appearance changes under different lighting sources. To improve efficiency and maintain consistency across each generated sequence, we fine-tune Stable Video Diffusion on robot execution videos for temporal lighting propagation. We evaluate our framework by measuring the visual quality of the generated sequences, assessing its effectiveness in improving the imitation learning policy performance (38.75\%) under six unseen real-world lighting conditions, and conduct ablation studies on individual modules of the proposed framework. We further showcase three downstream applications enabled by the proposed framework: background generation, object texture generation and distractor positioning. The code for the framework will be made publicly available.

Country of Origin
πŸ‡ΈπŸ‡ͺ Sweden

Page Count
9 pages

Category
Computer Science:
Robotics