Autoregressive Video Autoencoder with Decoupled Temporal and Spatial Context
By: Cuifeng Shen , Lumin Xu , Xingguo Zhu and more
Video autoencoders compress videos into compact latent representations for efficient reconstruction, playing a vital role in enhancing the quality and efficiency of video generation. However, existing video autoencoders often entangle spatial and temporal information, limiting their ability to capture temporal consistency and leading to suboptimal performance. To address this, we propose Autoregressive Video Autoencoder (ARVAE), which compresses and reconstructs each frame conditioned on its predecessor in an autoregressive manner, allowing flexible processing of videos with arbitrary lengths. ARVAE introduces a temporal-spatial decoupled representation that combines downsampled flow field for temporal coherence with spatial relative compensation for newly emerged content, achieving high compression efficiency without information loss. Specifically, the encoder compresses the current and previous frames into the temporal motion and spatial supplement, while the decoder reconstructs the original frame from the latent representations given the preceding frame. A multi-stage training strategy is employed to progressively optimize the model. Extensive experiments demonstrate that ARVAE achieves superior reconstruction quality with extremely lightweight models and small-scale training data. Moreover, evaluations on video generation tasks highlight its strong potential for downstream applications.
Similar Papers
DeCo-VAE: Learning Compact Latents for Video Reconstruction via Decoupled Representation
CV and Pattern Recognition
Makes videos smaller by separating key parts.
Hi-VAE: Efficient Video Autoencoding with Global and Detailed Motion
CV and Pattern Recognition
Makes videos smaller without losing quality.
VRAE: Vertical Residual Autoencoder for License Plate Denoising and Deblurring
CV and Pattern Recognition
Cleans blurry car pictures for better license plates.