Synthetic FMCW Radar Range Azimuth Maps Augmentation with Generative Diffusion Model
By: Zhaoze Wang , Changxu Zhang , Tai Fei and more
The scarcity and low diversity of well-annotated automotive radar datasets often limit the performance of deep-learning-based environmental perception. To overcome these challenges, we propose a conditional generative framework for synthesizing realistic Frequency-Modulated Continuous-Wave radar Range-Azimuth Maps. Our approach leverages a generative diffusion model to generate radar data for multiple object categories, including pedestrians, cars, and cyclists. Specifically, conditioning is achieved via Confidence Maps, where each channel represents a semantic class and encodes Gaussian-distributed annotations at target locations. To address radar-specific characteristics, we incorporate Geometry Aware Conditioning and Temporal Consistency Regularization into the generative process. Experiments on the ROD2021 dataset demonstrate that signal reconstruction quality improves by \SI{3.6}{dB} in Peak Signal-to-Noise Ratio over baseline methods, while training with a combination of real and synthetic datasets improves overall mean Average Precision by 4.15% compared with conventional image-processing-based augmentation. These results indicate that our generative framework not only produces physically plausible and diverse radar spectrum but also substantially improves model generalization in downstream tasks.
Similar Papers
Comprehensive Deployment-Oriented Assessment for Cross-Environment Generalization in Deep Learning-Based mmWave Radar Sensing
CV and Pattern Recognition
Radar sees people accurately in new places.
RadarGen: Automotive Radar Point Cloud Generation from Cameras
CV and Pattern Recognition
Makes self-driving cars see like real radar.
4D-RaDiff: Latent Diffusion for 4D Radar Point Cloud Generation
CV and Pattern Recognition
Makes self-driving cars see better in fog.