FLEG: Feed-Forward Language Embedded Gaussian Splatting from Any Views
By: Qijian Tian , Xin Tan , Jiayu Ying and more
We present FLEG, a feed-forward network that reconstructs language-embedded 3D Gaussians from any views. Previous straightforward solutions combine feed-forward reconstruction with Gaussian heads but suffer from fixed input views and insufficient 3D training data. In contrast, we propose a 3D-annotation-free training framework for 2D-to-3D lifting from arbitrary uncalibrated and unposed multi-view images. Since the framework does not require 3D annotations, we can leverage large-scale video data with easily obtained 2D instance information to enrich semantic embedding. We also propose an instance-guided contrastive learning to align 2D semantics with the 3D representations. In addition, to mitigate the high memory and computational cost of dense views, we further propose a geometry-semantic hierarchical sparsification strategy. Our FLEG efficiently reconstructs language-embedded 3D Gaussian representation in a feed-forward manner from arbitrary sparse or dense views, jointly producing accurate geometry, high-fidelity appearance, and language-aligned semantics. Extensive experiments show that it outperforms existing methods on various related tasks. Project page: https://fangzhou2000.github.io/projects/fleg.
Similar Papers
C3G: Learning Compact 3D Representations with 2K Gaussians
CV and Pattern Recognition
Builds detailed 3D worlds from few pictures.
Lang3D-XL: Language Embedded 3D Gaussians for Large-scale Scenes
CV and Pattern Recognition
Lets computers understand and change 3D worlds with words.
SemanticSplat: Feed-Forward 3D Scene Understanding with Language-Aware Gaussian Fields
CV and Pattern Recognition
Builds 3D worlds from a few pictures.