Score: 0

Multimodal and Multiview Deep Fusion for Autonomous Marine Navigation

Published: May 2, 2025 | arXiv ID: 2505.01615v1

By: Dimitrios Dagdilelis, Panagiotis Grigoriadis, Roberto Galeazzi

Potential Business Impact:

Helps boats see better in fog and storms.

Business Areas:
Autonomous Vehicles Transportation

We propose a cross attention transformer based method for multimodal sensor fusion to build a birds eye view of a vessels surroundings supporting safer autonomous marine navigation. The model deeply fuses multiview RGB and long wave infrared images with sparse LiDAR point clouds. Training also integrates X band radar and electronic chart data to inform predictions. The resulting view provides a detailed reliable scene representation improving navigational accuracy and robustness. Real world sea trials confirm the methods effectiveness even in adverse weather and complex maritime settings.

Page Count
14 pages

Category
Computer Science:
CV and Pattern Recognition