Data Cartography for Detecting Memorization Hotspots and Guiding Data Interventions in Generative Models
By: Laksh Patel, Neel Shanbhag
Potential Business Impact:
Makes AI forget private data it learned.
Modern generative models risk overfitting and unintentionally memorizing rare training examples, which can be extracted by adversaries or inflate benchmark performance. We propose Generative Data Cartography (GenDataCarto), a data-centric framework that assigns each pretraining sample a difficulty score (early-epoch loss) and a memorization score (frequency of ``forget events''), then partitions examples into four quadrants to guide targeted pruning and up-/down-weighting. We prove that our memorization score lower-bounds classical influence under smoothness assumptions and that down-weighting high-memorization hotspots provably decreases the generalization gap via uniform stability bounds. Empirically, GenDataCarto reduces synthetic canary extraction success by over 40\% at just 10\% data pruning, while increasing validation perplexity by less than 0.5\%. These results demonstrate that principled data interventions can dramatically mitigate leakage with minimal cost to generative performance.
Similar Papers
Memorization in 3D Shape Generation: An Empirical Study
CV and Pattern Recognition
Finds if AI copies 3D shapes it learned.
Unconsciously Forget: Mitigating Memorization; Without Knowing What is being Memorized
CV and Pattern Recognition
Stops AI from copying art it learned from.
Beyond Memorization: Gradient Projection Enables Selective Learning in Diffusion Models
Machine Learning (CS)
Stops AI from copying private images.