Data Unlearning Beyond Uniform Forgetting via Diffusion Time and Frequency Selection
By: Jinseong Park, Mijung Park
Potential Business Impact:
Removes unwanted images from AI art.
Data unlearning aims to remove the influence of specific training samples from a trained model without requiring full retraining. Unlike concept unlearning, data unlearning in diffusion models remains underexplored and often suffers from quality degradation or incomplete forgetting. To address this, we first observe that most existing methods attempt to unlearn the samples at all diffusion time steps equally, leading to poor-quality generation. We argue that forgetting occurs disproportionately across time and frequency, depending on the model and scenarios. By selectively focusing on specific time-frequency ranges during training, we achieve samples with higher aesthetic quality and lower noise. We validate this improvement by applying our time-frequency selective approach to diverse settings, including gradient-based and preference optimization objectives, as well as both image-level and text-to-image tasks. Finally, to evaluate both deletion and quality of unlearned data samples, we propose a simple normalized version of SSCD. Together, our analysis and methods establish a clearer understanding of the unique challenges in data unlearning for diffusion models, providing practical strategies to improve both evaluation and unlearning performance.
Similar Papers
Data Unlearning in Diffusion Models
Machine Learning (CS)
Removes unwanted images from AI art generators.
Distill, Forget, Repeat: A Framework for Continual Unlearning in Text-to-Image Diffusion Models
Machine Learning (CS)
Removes unwanted data from AI without retraining.
Not All Data Are Unlearned Equally
Computation and Language
Removes unwanted information from AI minds.