StableMotion: Repurposing Diffusion-Based Image Priors for Motion Estimation
By: Ziyi Wang , Haipeng Li , Lin Sui and more
Potential Business Impact:
Fixes wobbly or distorted pictures using AI.
We present StableMotion, a novel framework leverages knowledge (geometry and content priors) from pretrained large-scale image diffusion models to perform motion estimation, solving single-image-based image rectification tasks such as Stitched Image Rectangling (SIR) and Rolling Shutter Correction (RSC). Specifically, StableMotion framework takes text-to-image Stable Diffusion (SD) models as backbone and repurposes it into an image-to-motion estimator. To mitigate inconsistent output produced by diffusion models, we propose Adaptive Ensemble Strategy (AES) that consolidates multiple outputs into a cohesive, high-fidelity result. Additionally, we present the concept of Sampling Steps Disaster (SSD), the counterintuitive scenario where increasing the number of sampling steps can lead to poorer outcomes, which enables our framework to achieve one-step inference. StableMotion is verified on two image rectification tasks and delivers state-of-the-art performance in both, as well as showing strong generalizability. Supported by SSD, StableMotion offers a speedup of 200 times compared to previous diffusion model-based methods.
Similar Papers
SD-Acc: Accelerating Stable Diffusion through Phase-aware Sampling and Hardware Co-Optimizations
Hardware Architecture
Makes AI art generators faster and use less power.
Rethinking Video Super-Resolution: Towards Diffusion-Based Methods without Motion Alignment
CV and Pattern Recognition
Makes blurry videos super clear without extra work.
Taming Stable Diffusion for Computed Tomography Blind Super-Resolution
Image and Video Processing
Makes X-rays clearer with less radiation.