Score: 0

Adaptive Domain Shift in Diffusion Models for Cross-Modality Image Translation

Published: January 26, 2026 | arXiv ID: 2601.18623v1

By: Zihao Wang, Yuzhou Chen, Shaogang Ren

Potential Business Impact:

Makes pictures change into other pictures better.

Business Areas:
Visual Search Internet Services

Cross-modal image translation remains brittle and inefficient. Standard diffusion approaches often rely on a single, global linear transfer between domains. We find that this shortcut forces the sampler to traverse off-manifold, high-cost regions, inflating the correction burden and inviting semantic drift. We refer to this shared failure mode as fixed-schedule domain transfer. In this paper, we embed domain-shift dynamics directly into the generative process. Our model predicts a spatially varying mixing field at every reverse step and injects an explicit, target-consistent restoration term into the drift. This in-step guidance keeps large updates on-manifold and shifts the model's role from global alignment to local residual correction. We provide a continuous-time formulation with an exact solution form and derive a practical first-order sampler that preserves marginal consistency. Empirically, across translation tasks in medical imaging, remote sensing, and electroluminescence semantic mapping, our framework improves structural fidelity and semantic consistency while converging in fewer denoising steps.

Page Count
31 pages

Category
Computer Science:
CV and Pattern Recognition