Identifying Memorization of Diffusion Models through p-Laplace Analysis
By: Jonathan Brokman , Amit Giloni , Omer Hofman and more
Potential Business Impact:
Finds if AI copied its training pictures.
Diffusion models, today's leading image generative models, estimate the score function, i.e. the gradient of the log probability of (perturbed) data samples, without direct access to the underlying probability distribution. This work investigates whether the estimated score function can be leveraged to compute higher-order differentials, namely p-Laplace operators. We show here these operators can be employed to identify memorized training data. We propose a numerical p-Laplace approximation based on the learned score functions, showing its effectiveness in identifying key features of the probability landscape. We analyze the structured case of Gaussian mixture models, and demonstrate the results carry-over to image generative models, where memorization identification based on the p-Laplace operator is performed for the first time.
Similar Papers
Malliavin Calculus for Score-based Diffusion Models
Machine Learning (CS)
Makes AI create realistic images and sounds.
Laplacian Score Sharpening for Mitigating Hallucination in Diffusion Models
CV and Pattern Recognition
Fixes AI art so it looks real.
Quantifying the Ease of Reproducing Training Data in Unconditional Diffusion Models
Machine Learning (CS)
Finds and fixes art that's too easy to copy.