Towards Efficient Low-rate Image Compression with Frequency-aware Diffusion Prior Refinement
By: Yichong Xia , Yimin Zhou , Jinpeng Wang and more
Recent advancements in diffusion-based generative priors have enabled visually plausible image compression at extremely low bit rates. However, existing approaches suffer from slow sampling processes and suboptimal bit allocation due to fragmented training paradigms. In this work, we propose Accelerate \textbf{Diff}usion-based Image Compression via \textbf{C}onsistency Prior \textbf{R}efinement (DiffCR), a novel compression framework for efficient and high-fidelity image reconstruction. At the heart of DiffCR is a Frequency-aware Skip Estimation (FaSE) module that refines the $ε$-prediction prior from a pre-trained latent diffusion model and aligns it with compressed latents at different timesteps via Frequency Decoupling Attention (FDA). Furthermore, a lightweight consistency estimator enables fast \textbf{two-step decoding} by preserving the semantic trajectory of diffusion sampling. Without updating the backbone diffusion model, DiffCR achieves substantial bitrate savings (27.2\% BD-rate (LPIPS) and 65.1\% BD-rate (PSNR)) and over $10\times$ speed-up compared to SOTA diffusion-based compression baselines.
Similar Papers
Towards Facial Image Compression with Consistency Preserving Diffusion Prior
CV and Pattern Recognition
Makes tiny face pictures look good and clear.
Generative Image Coding with Diffusion Prior
CV and Pattern Recognition
Makes pictures look good even when squeezed small.
Generative Preprocessing for Image Compression with Pre-trained Diffusion Models
Image and Video Processing
Makes pictures look better when they're made smaller.