Compressing image encoders via latent distillation
By: Caroline Mazini Rodrigues, Nicolas Keriven, Thomas Maugey
Potential Business Impact:
Makes AI image tools work on small devices.
Deep learning models for image compression often face practical limitations in hardware-constrained applications. Although these models achieve high-quality reconstructions, they are typically complex, heavyweight, and require substantial training data and computational resources. We propose a methodology to partially compress these networks by reducing the size of their encoders. Our approach uses a simplified knowledge distillation strategy to approximate the latent space of the original models with less data and shorter training, yielding lightweight encoders from heavyweight ones. We evaluate the resulting lightweight encoders across two different architectures on the image compression task. Experiments show that our method preserves reconstruction quality and statistical fidelity better than training lightweight encoders with the original loss, making it practical for resource-limited environments.
Similar Papers
Efficient Learned Image Compression Through Knowledge Distillation
CV and Pattern Recognition
Makes AI image compression faster and use less power.
Both Semantics and Reconstruction Matter: Making Representation Encoders Ready for Text-to-Image Generation and Editing
CV and Pattern Recognition
Makes AI create better, more detailed pictures.
Higher fidelity perceptual image and video compression with a latent conditioned residual denoising diffusion model
Image and Video Processing
Makes pictures look good while keeping details.