Who Made This? Fake Detection and Source Attribution with Diffusion Features
By: Simone Bonechi, Paolo Andreini, Barbara Toniella Corradini
Potential Business Impact:
Finds fake pictures made by computers.
The rapid progress of generative diffusion models has enabled the creation of synthetic images that are increasingly difficult to distinguish from real ones, raising concerns about authenticity, copyright, and misinformation. Existing supervised detectors often struggle to generalize across unseen generators, requiring extensive labeled data and frequent retraining. We introduce FRIDA (Fake-image Recognition and source Identification via Diffusion-features Analysis), a lightweight framework that leverages internal activations from a pre-trained diffusion model for deepfake detection and source generator attribution. A k-nearest-neighbor classifier applied to diffusion features achieves state-of-the-art cross-generator performance without fine-tuning, while a compact neural model enables accurate source attribution. These results show that diffusion representations inherently encode generator-specific patterns, providing a simple and interpretable foundation for synthetic image forensics.
Similar Papers
DiffusionFF: Face Forgery Detection via Diffusion-based Artifact Localization
CV and Pattern Recognition
Finds fake faces by spotting tiny digital mistakes.
A Hybrid Deep Learning and Forensic Approach for Robust Deepfake Detection
CV and Pattern Recognition
Finds fake videos by combining clues.
Explore the vulnerability of black-box models via diffusion models
CV and Pattern Recognition
Steals AI art to trick other AIs.