LiteVPNet: A Lightweight Network for Video Encoding Control in Quality-Critical Applications
By: Vibhoothi Vibhoothi, François Pitié, Anil Kokaram
Potential Business Impact:
Makes movies stream better and use less power.
In the last decade, video workflows in the cinema production ecosystem have presented new use cases for video streaming technology. These new workflows, e.g. in On-set Virtual Production, present the challenge of requiring precise quality control and energy efficiency. Existing approaches to transcoding often fall short of these requirements, either due to a lack of quality control or computational overhead. To fill this gap, we present a lightweight neural network (LiteVPNet) for accurately predicting Quantisation Parameters for NVENC AV1 encoders that achieve a specified VMAF score. We use low-complexity features, including bitstream characteristics, video complexity measures, and CLIP-based semantic embeddings. Our results demonstrate that LiteVPNet achieves mean VMAF errors below 1.2 points across a wide range of quality targets. Notably, LiteVPNet achieves VMAF errors within 2 points for over 87% of our test corpus, c.f. approx 61% with state-of-the-art methods. LiteVPNet's performance across various quality regions highlights its applicability for enhancing high-value content transport and streaming for more energy-efficient, high-quality media experiences.
Similar Papers
Predicting Encoding Energy from Low-Pass Anchors for Green Video Streaming
Multimedia
Saves energy watching videos online.
LeanVAE: An Ultra-Efficient Reconstruction VAE for Video Diffusion Models
CV and Pattern Recognition
Makes video creation much faster and cheaper.
Less is More: Token-Efficient Video-QA via Adaptive Frame-Pruning and Semantic Graph Integration
CV and Pattern Recognition
Makes videos easier for computers to understand.