Score: 3

SE360: Semantic Edit in 360$^\circ$ Panoramas via Hierarchical Data Construction

Published: December 23, 2025 | arXiv ID: 2512.19943v1

By: Haoyi Zhong , Fang-Lue Zhang , Andrew Chalmers and more

Potential Business Impact:

Edits 360 photos with simple text instructions.

Business Areas:
Image Recognition Data and Analytics, Software

While instruction-based image editing is emerging, extending it to 360$^\circ$ panoramas introduces additional challenges. Existing methods often produce implausible results in both equirectangular projections (ERP) and perspective views. To address these limitations, we propose SE360, a novel framework for multi-condition guided object editing in 360$^\circ$ panoramas. At its core is a novel coarse-to-fine autonomous data generation pipeline without manual intervention. This pipeline leverages a Vision-Language Model (VLM) and adaptive projection adjustment for hierarchical analysis, ensuring the holistic segmentation of objects and their physical context. The resulting data pairs are both semantically meaningful and geometrically consistent, even when sourced from unlabeled panoramas. Furthermore, we introduce a cost-effective, two-stage data refinement strategy to improve data realism and mitigate model overfitting to erase artifacts. Based on the constructed dataset, we train a Transformer-based diffusion model to allow flexible object editing guided by text, mask, or reference image in 360$^\circ$ panoramas. Our experiments demonstrate that our method outperforms existing methods in both visual quality and semantic accuracy.

Country of Origin
🇳🇿 🇦🇺 New Zealand, Australia

Repos / Data Links

Page Count
26 pages

Category
Computer Science:
CV and Pattern Recognition