CAPTAIN: Semantic Feature Injection for Memorization Mitigation in Text-to-Image Diffusion Models
By: Tong Zhang, Carlos Hinojosa, Bernard Ghanem
Potential Business Impact:
Stops AI from copying its training pictures.
Diffusion models can unintentionally reproduce training examples, raising privacy and copyright concerns as these systems are increasingly deployed at scale. Existing inference-time mitigation methods typically manipulate classifier-free guidance (CFG) or perturb prompt embeddings; however, they often struggle to reduce memorization without compromising alignment with the conditioning prompt. We introduce CAPTAIN, a training-free framework that mitigates memorization by directly modifying latent features during denoising. CAPTAIN first applies frequency-based noise initialization to reduce the tendency to replicate memorized patterns early in the denoising process. It then identifies the optimal denoising timesteps for feature injection and localizes memorized regions. Finally, CAPTAIN injects semantically aligned features from non-memorized reference images into localized latent regions, suppressing memorization while preserving prompt fidelity and visual quality. Our experiments show that CAPTAIN achieves substantial reductions in memorization compared to CFG-based baselines while maintaining strong alignment with the intended prompt.
Similar Papers
Unconsciously Forget: Mitigating Memorization; Without Knowing What is being Memorized
CV and Pattern Recognition
Stops AI from copying art it learned from.
DiffCAP: Diffusion-based Cumulative Adversarial Purification for Vision Language Models
CV and Pattern Recognition
Fixes AI mistakes caused by tricky images.
Beyond the Noise: Aligning Prompts with Latent Representations in Diffusion Models
CV and Pattern Recognition
Finds bad AI pictures while they're still being made.