Bayesian Monocular Depth Refinement via Neural Radiance Fields
By: Arun Muthukkumar
Potential Business Impact:
Makes 3D pictures show tiny details better.
Monocular depth estimation has applications in many fields, such as autonomous navigation and extended reality, making it an essential computer vision task. However, current methods often produce smooth depth maps that lack the fine geometric detail needed for accurate scene understanding. We propose MDENeRF, an iterative framework that refines monocular depth estimates using depth information from Neural Radiance Fields (NeRFs). MDENeRF consists of three components: (1) an initial monocular estimate for global structure, (2) a NeRF trained on perturbed viewpoints, with per-pixel uncertainty, and (3) Bayesian fusion of the noisy monocular and NeRF depths. We derive NeRF uncertainty from the volume rendering process to iteratively inject high-frequency fine details. Meanwhile, our monocular prior maintains global structure. We demonstrate superior performance on key metrics and experiments using indoor scenes from the SUN RGB-D dataset.
Similar Papers
Joint Learning of Depth, Pose, and Local Radiance Field for Large Scale Monocular 3D Reconstruction
CV and Pattern Recognition
Creates 3D worlds from one camera.
Improving Geometric Consistency for 360-Degree Neural Radiance Fields in Indoor Scenarios
CV and Pattern Recognition
Makes computer pictures of rooms look more real.
NeRFs are Mirror Detectors: Using Structural Similarity for Multi-View Mirror Scene Reconstruction with 3D Surface Primitives
CV and Pattern Recognition
Makes computer pictures show reflections correctly.