Score: 0

Regularized Personalization of Text-to-Image Diffusion Models without Distributional Drift

Published: May 26, 2025 | arXiv ID: 2505.19519v2

By: Gihoon Kim, Hyungjin Park, Taesup Kim

Potential Business Impact:

Teaches computers to draw new things without forgetting.

Business Areas:
Personalization Commerce and Shopping

Personalization using text-to-image diffusion models involves adapting a pretrained model to novel subjects with only a few image examples. This task presents a fundamental challenge, as the model must not only learn the new subject effectively but also preserve its ability to generate diverse and coherent outputs across a wide range of prompts. In other words, successful personalization requires integrating new concepts without forgetting previously learned generative capabilities. Forgetting denotes unintended distributional drift, where the model's output distribution deviates from that of the original pretrained model. In this paper, we provide an analysis of this issue and identify a mismatch between standard training objectives and the goals of personalization. To address this, we propose a new training objective based on a Lipschitz-bounded formulation that explicitly constrains deviation from the pretrained distribution. Our method provides improved control over distributional drift and performs well even in data-scarce scenarios. Experimental results demonstrate that our approach consistently outperforms existing personalization methods, achieving higher CLIP-T, CLIP-I, and DINO scores.

Page Count
22 pages

Category
Computer Science:
CV and Pattern Recognition