CGVQM+D: Computer Graphics Video Quality Metric and Dataset
By: Akshay Jindal , Nabil Sadaka , Manu Mathew Thomas and more
Potential Business Impact:
Makes computer-generated videos look more real.
While existing video and image quality datasets have extensively studied natural videos and traditional distortions, the perception of synthetic content and modern rendering artifacts remains underexplored. We present a novel video quality dataset focused on distortions introduced by advanced rendering techniques, including neural supersampling, novel-view synthesis, path tracing, neural denoising, frame interpolation, and variable rate shading. Our evaluations show that existing full-reference quality metrics perform sub-optimally on these distortions, with a maximum Pearson correlation of 0.78. Additionally, we find that the feature space of pre-trained 3D CNNs aligns strongly with human perception of visual quality. We propose CGVQM, a full-reference video quality metric that significantly outperforms existing metrics while generating both per-pixel error maps and global quality scores. Our dataset and metric implementation is available at https://github.com/IntelLabs/CGVQM.
Similar Papers
No-Reference Rendered Video Quality Assessment: Dataset and Metrics
CV and Pattern Recognition
Makes computer-made videos look better.
SDQM: Synthetic Data Quality Metric for Object Detection Dataset Evaluation
CV and Pattern Recognition
Checks fake pictures for computers to learn better.
Semantically-Aware Game Image Quality Assessment
CV and Pattern Recognition
Makes video game graphics look better automatically.