Guided Model-based LiDAR Super-Resolution for Resource-Efficient Automotive scene Segmentation
By: Alexandros Gkillas, Nikos Piperigkos, Aris S. Lalos
Potential Business Impact:
Makes cheap car sensors see like expensive ones.
High-resolution LiDAR data plays a critical role in 3D semantic segmentation for autonomous driving, but the high cost of advanced sensors limits large-scale deployment. In contrast, low-cost sensors such as 16-channel LiDAR produce sparse point clouds that degrade segmentation accuracy. To overcome this, we introduce the first end-to-end framework that jointly addresses LiDAR super-resolution (SR) and semantic segmentation. The framework employs joint optimization during training, allowing the SR module to incorporate semantic cues and preserve fine details, particularly for smaller object classes. A new SR loss function further directs the network to focus on regions of interest. The proposed lightweight, model-based SR architecture uses significantly fewer parameters than existing LiDAR SR approaches, while remaining easily compatible with segmentation networks. Experiments show that our method achieves segmentation performance comparable to models operating on high-resolution and costly 64-channel LiDAR data.
Similar Papers
Real Time Semantic Segmentation of High Resolution Automotive LiDAR Scans
Robotics
Helps self-driving cars see better in real-time.
Semantic Segmentation Algorithm Based on Light Field and LiDAR Fusion
CV and Pattern Recognition
Helps self-driving cars see through obstacles.
Boosting LiDAR-Based Localization with Semantic Insight: Camera Projection versus Direct LiDAR Segmentation
Robotics
Helps self-driving cars see better with cameras and lasers.