FlyAwareV2: A Multimodal Cross-Domain UAV Dataset for Urban Scene Understanding
By: Francesco Barbato, Matteo Caligiuri, Pietro Zanuttigh
Potential Business Impact:
Helps drones see cities better in all weather.
The development of computer vision algorithms for Unmanned Aerial Vehicle (UAV) applications in urban environments heavily relies on the availability of large-scale datasets with accurate annotations. However, collecting and annotating real-world UAV data is extremely challenging and costly. To address this limitation, we present FlyAwareV2, a novel multimodal dataset encompassing both real and synthetic UAV imagery tailored for urban scene understanding tasks. Building upon the recently introduced SynDrone and FlyAware datasets, FlyAwareV2 introduces several new key contributions: 1) Multimodal data (RGB, depth, semantic labels) across diverse environmental conditions including varying weather and daytime; 2) Depth maps for real samples computed via state-of-the-art monocular depth estimation; 3) Benchmarks for RGB and multimodal semantic segmentation on standard architectures; 4) Studies on synthetic-to-real domain adaptation to assess the generalization capabilities of models trained on the synthetic data. With its rich set of annotations and environmental diversity, FlyAwareV2 provides a valuable resource for research on UAV-based 3D urban scene understanding.
Similar Papers
UAV-MM3D: A Large-Scale Synthetic Benchmark for 3D Perception of Unmanned Aerial Vehicles with Multi-Modal Data
CV and Pattern Recognition
Creates realistic drone videos for training AI.
Aerial-ground Cross-modal Localization: Dataset, Ground-truth, and Benchmark
Robotics
Helps robots find their way using 3D maps.
Crossmodal learning for Crop Canopy Trait Estimation
CV and Pattern Recognition
Makes satellite pictures show tiny plant details.