Adaptive Domain Shift in Diffusion Models for Cross-Modality Image Translation
By: Zihao Wang, Yuzhou Chen, Shaogang Ren
Potential Business Impact:
Makes pictures change into other pictures better.
Cross-modal image translation remains brittle and inefficient. Standard diffusion approaches often rely on a single, global linear transfer between domains. We find that this shortcut forces the sampler to traverse off-manifold, high-cost regions, inflating the correction burden and inviting semantic drift. We refer to this shared failure mode as fixed-schedule domain transfer. In this paper, we embed domain-shift dynamics directly into the generative process. Our model predicts a spatially varying mixing field at every reverse step and injects an explicit, target-consistent restoration term into the drift. This in-step guidance keeps large updates on-manifold and shifts the model's role from global alignment to local residual correction. We provide a continuous-time formulation with an exact solution form and derive a practical first-order sampler that preserves marginal consistency. Empirically, across translation tasks in medical imaging, remote sensing, and electroluminescence semantic mapping, our framework improves structural fidelity and semantic consistency while converging in fewer denoising steps.
Similar Papers
Test-Time Modification: Inverse Domain Transformation for Robust Perception
CV and Pattern Recognition
Makes AI see in new places without retraining.
CycleDiff: Cycle Diffusion Models for Unpaired Image-to-image Translation
CV and Pattern Recognition
Changes pictures from one style to another.
Semi-Supervised Domain Adaptation with Latent Diffusion for Pathology Image Classification
CV and Pattern Recognition
Helps AI see cancer better in different hospitals.