Regularized Personalization of Text-to-Image Diffusion Models without Distributional Drift
By: Gihoon Kim, Hyungjin Park, Taesup Kim
Potential Business Impact:
Teaches computers to draw new things without forgetting.
Personalization using text-to-image diffusion models involves adapting a pretrained model to novel subjects with only a few image examples. This task presents a fundamental challenge, as the model must not only learn the new subject effectively but also preserve its ability to generate diverse and coherent outputs across a wide range of prompts. In other words, successful personalization requires integrating new concepts without forgetting previously learned generative capabilities. Forgetting denotes unintended distributional drift, where the model's output distribution deviates from that of the original pretrained model. In this paper, we provide an analysis of this issue and identify a mismatch between standard training objectives and the goals of personalization. To address this, we propose a new training objective based on a Lipschitz-bounded formulation that explicitly constrains deviation from the pretrained distribution. Our method provides improved control over distributional drift and performs well even in data-scarce scenarios. Experimental results demonstrate that our approach consistently outperforms existing personalization methods, achieving higher CLIP-T, CLIP-I, and DINO scores.
Similar Papers
LLM-Enabled Style and Content Regularization for Personalized Text-to-Image Generation
CV and Pattern Recognition
Makes AI pictures match your style better.
Energy-Guided Optimization for Personalized Image Editing with Pretrained Text-to-Image Diffusion Models
CV and Pattern Recognition
Changes pictures to match your exact ideas.
LUSD: Localized Update Score Distillation for Text-Guided Image Editing
Graphics
Makes AI better at changing pictures without messing up.