Have We Scene It All? Scene Graph-Aware Deep Point Cloud Compression
By: Nikolaos Stathoulopoulos, Christoforos Kanellakis, George Nikolakopoulos
Potential Business Impact:
Shrinks 3D robot data, keeping details for better teamwork.
Efficient transmission of 3D point cloud data is critical for advanced perception in centralized and decentralized multi-agent robotic systems, especially nowadays with the growing reliance on edge and cloud-based processing. However, the large and complex nature of point clouds creates challenges under bandwidth constraints and intermittent connectivity, often degrading system performance. We propose a deep compression framework based on semantic scene graphs. The method decomposes point clouds into semantically coherent patches and encodes them into compact latent representations with semantic-aware encoders conditioned by Feature-wise Linear Modulation (FiLM). A folding-based decoder, guided by latent features and graph node attributes, enables structurally accurate reconstruction. Experiments on the SemanticKITTI and nuScenes datasets show that the framework achieves state-of-the-art compression rates, reducing data size by up to 98% while preserving both structural and semantic fidelity. In addition, it supports downstream applications such as multi-robot pose graph optimization and map merging, achieving trajectory accuracy and map alignment comparable to those obtained with raw LiDAR scans.
Similar Papers
Transmit Weights, Not Features: Orthogonal-Basis Aided Wireless Point-Cloud Transmission
Machine Learning (CS)
Sends 3D shapes over Wi-Fi more efficiently.
Point-Plane Projections for Accurate LiDAR Semantic Segmentation in Small Data Scenarios
CV and Pattern Recognition
Helps self-driving cars see better with less data.
ROI-Guided Point Cloud Geometry Compression Towards Human and Machine Vision
CV and Pattern Recognition
Shrinks 3D scans without losing important details.