Spherical Vision Transformers for Audio-Visual Saliency Prediction in 360-Degree Videos
By: Mert Cokelek , Halit Ozsoy , Nevrez Imamoglu and more
Potential Business Impact:
Makes VR videos show where people look.
Omnidirectional videos (ODVs) are redefining viewer experiences in virtual reality (VR) by offering an unprecedented full field-of-view (FOV). This study extends the domain of saliency prediction to 360-degree environments, addressing the complexities of spherical distortion and the integration of spatial audio. Contextually, ODVs have transformed user experience by adding a spatial audio dimension that aligns sound direction with the viewer's perspective in spherical scenes. Motivated by the lack of comprehensive datasets for 360-degree audio-visual saliency prediction, our study curates YT360-EyeTracking, a new dataset of 81 ODVs, each observed under varying audio-visual conditions. Our goal is to explore how to utilize audio-visual cues to effectively predict visual saliency in 360-degree videos. Towards this aim, we propose two novel saliency prediction models: SalViT360, a vision-transformer-based framework for ODVs equipped with spherical geometry-aware spatio-temporal attention layers, and SalViT360-AV, which further incorporates transformer adapters conditioned on audio input. Our results on a number of benchmark datasets, including our YT360-EyeTracking, demonstrate that SalViT360 and SalViT360-AV significantly outperform existing methods in predicting viewer attention in 360-degree scenes. Interpreting these results, we suggest that integrating spatial audio cues in the model architecture is crucial for accurate saliency prediction in omnidirectional videos. Code and dataset will be available at https://cyberiada.github.io/SalViT360.
Similar Papers
Sphere-GAN: a GAN-based Approach for Saliency Estimation in 360° Videos
CV and Pattern Recognition
Finds important parts of 360-degree videos.
OmniAudio: Generating Spatial Audio from 360-Degree Video
Audio and Speech Processing
Makes 360 videos sound like you're there.
Controllable Audio-Visual Viewpoint Generation from 360° Spatial Information
Multimedia
Makes videos sound and look right from any angle.