VisLanding: Monocular 3D Perception for UAV Safe Landing via Depth-Normal Synergy
By: Zhuoyue Tan , Boyong He , Yuxiang Ji and more
Potential Business Impact:
Helps drones land safely anywhere, even unknown places.
This paper presents VisLanding, a monocular 3D perception-based framework for safe UAV (Unmanned Aerial Vehicle) landing. Addressing the core challenge of autonomous UAV landing in complex and unknown environments, this study innovatively leverages the depth-normal synergy prediction capabilities of the Metric3D V2 model to construct an end-to-end safe landing zones (SLZ) estimation framework. By introducing a safe zone segmentation branch, we transform the landing zone estimation task into a binary semantic segmentation problem. The model is fine-tuned and annotated using the WildUAV dataset from a UAV perspective, while a cross-domain evaluation dataset is constructed to validate the model's robustness. Experimental results demonstrate that VisLanding significantly enhances the accuracy of safe zone identification through a depth-normal joint optimization mechanism, while retaining the zero-shot generalization advantages of Metric3D V2. The proposed method exhibits superior generalization and robustness in cross-domain testing compared to other approaches. Furthermore, it enables the estimation of landing zone area by integrating predicted depth and normal information, providing critical decision-making support for practical applications.
Similar Papers
Vision-Based Risk Aware Emergency Landing for UAVs in Complex Urban Environments
Robotics
Drones find safe spots to land in busy cities.
Reinforcement Learning-Based Monocular Vision Approach for Autonomous UAV Landing
Robotics
Drones land safely using just one camera.
Zero-Shot Metric Depth Estimation via Monocular Visual-Inertial Rescaling for Autonomous Aerial Navigation
Robotics
Helps drones see how far things are.