Multimodal and Multiview Deep Fusion for Autonomous Marine Navigation
By: Dimitrios Dagdilelis, Panagiotis Grigoriadis, Roberto Galeazzi
Potential Business Impact:
Helps boats see better in fog and storms.
We propose a cross attention transformer based method for multimodal sensor fusion to build a birds eye view of a vessels surroundings supporting safer autonomous marine navigation. The model deeply fuses multiview RGB and long wave infrared images with sparse LiDAR point clouds. Training also integrates X band radar and electronic chart data to inform predictions. The resulting view provides a detailed reliable scene representation improving navigational accuracy and robustness. Real world sea trials confirm the methods effectiveness even in adverse weather and complex maritime settings.
Similar Papers
Deep Learning-Based Multi-Modal Fusion for Robust Robot Perception and Navigation
Machine Learning (CS)
Helps robots see and move better in tricky places.
Lightweight Multimodal Artificial Intelligence Framework for Maritime Multi-Scene Recognition
CV and Pattern Recognition
Helps robots understand ocean pictures better.
Bridging Spectral-wise and Multi-spectral Depth Estimation via Geometry-guided Contrastive Learning
CV and Pattern Recognition
Helps self-driving cars see better in any weather.