Sphere-GAN: a GAN-based Approach for Saliency Estimation in 360° Videos
By: Mahmoud Z. A. Wahba, Sara Baldoni, Federica Battisti
Potential Business Impact:
Finds important parts of 360-degree videos.
The recent success of immersive applications is pushing the research community to define new approaches to process 360{\deg} images and videos and optimize their transmission. Among these, saliency estimation provides a powerful tool that can be used to identify visually relevant areas and, consequently, adapt processing algorithms. Although saliency estimation has been widely investigated for 2D content, very few algorithms have been proposed for 360{\deg} saliency estimation. Towards this goal, we introduce Sphere-GAN, a saliency detection model for 360{\deg} videos that leverages a Generative Adversarial Network with spherical convolutions. Extensive experiments were conducted using a public 360{\deg} video saliency dataset, and the results demonstrate that Sphere-GAN outperforms state-of-the-art models in accurately predicting saliency maps.
Similar Papers
Spherical Vision Transformers for Audio-Visual Saliency Prediction in 360-Degree Videos
CV and Pattern Recognition
Makes VR videos show where people look.
Seam360GS: Seamless 360° Gaussian Splatting from Real-World Omnidirectional Images
CV and Pattern Recognition
Fixes blurry 360-degree pictures for VR.
Physically Aware 360$^\circ$ View Generation from a Single Image using Disentangled Scene Embeddings
CV and Pattern Recognition
Creates realistic 3D views from one picture.