Score: 2

Diffusion-Based Restoration for Multi-Modal 3D Object Detection in Adverse Weather

Published: December 15, 2025 | arXiv ID: 2512.13107v1

By: Zhijian He , Feifei Liu , Yuwei Li and more

Potential Business Impact:

Helps self-driving cars see better in bad weather.

Business Areas:
Image Recognition Data and Analytics, Software

Multi-modal 3D object detection is important for reliable perception in robotics and autonomous driving. However, its effectiveness remains limited under adverse weather conditions due to weather-induced distortions and misalignment between different data modalities. In this work, we propose DiffFusion, a novel framework designed to enhance robustness in challenging weather through diffusion-based restoration and adaptive cross-modal fusion. Our key insight is that diffusion models possess strong capabilities for denoising and generating data that can adapt to various weather conditions. Building on this, DiffFusion introduces Diffusion-IR restoring images degraded by weather effects and Point Cloud Restoration (PCR) compensating for corrupted LiDAR data using image object cues. To tackle misalignments between two modalities, we develop Bidirectional Adaptive Fusion and Alignment Module (BAFAM). It enables dynamic multi-modal fusion and bidirectional bird's-eye view (BEV) alignment to maintain consistent spatial correspondence. Extensive experiments on three public datasets show that DiffFusion achieves state-of-the-art robustness under adverse weather while preserving strong clean-data performance. Zero-shot results on the real-world DENSE dataset further validate its generalization. The implementation of our DiffFusion will be released as open-source.

Country of Origin
🇨🇳 🇭🇰 China, Hong Kong

Page Count
9 pages

Category
Computer Science:
CV and Pattern Recognition