Score: 1

Feed-Forward 3D Gaussian Splatting Compression with Long-Context Modeling

Published: November 30, 2025 | arXiv ID: 2512.00877v1

By: Zhening Liu , Rui Song , Yushi Huang and more

Potential Business Impact:

Shrinks 3D scenes to fit on phones.

Business Areas:
Semantic Search Internet Services

3D Gaussian Splatting (3DGS) has emerged as a revolutionary 3D representation. However, its substantial data size poses a major barrier to widespread adoption. While feed-forward 3DGS compression offers a practical alternative to costly per-scene per-train compressors, existing methods struggle to model long-range spatial dependencies, due to the limited receptive field of transform coding networks and the inadequate context capacity in entropy models. In this work, we propose a novel feed-forward 3DGS compression framework that effectively models long-range correlations to enable highly compact and generalizable 3D representations. Central to our approach is a large-scale context structure that comprises thousands of Gaussians based on Morton serialization. We then design a fine-grained space-channel auto-regressive entropy model to fully leverage this expansive context. Furthermore, we develop an attention-based transform coding model to extract informative latent priors by aggregating features from a wide range of neighboring Gaussians. Our method yields a $20\times$ compression ratio for 3DGS in a feed-forward inference and achieves state-of-the-art performance among generalizable codecs.

Country of Origin
🇭🇰 Hong Kong

Page Count
17 pages

Category
Computer Science:
CV and Pattern Recognition