PRISM: Precision-Recall Informed Data-Free Knowledge Distillation via Generative Diffusion
By: Xuewan He , Jielei Wang , Zihan Cheng and more
Potential Business Impact:
Makes AI learn from fake pictures, not real ones.
Data-free knowledge distillation (DFKD) transfers knowledge from a teacher to a student without access to the real in-distribution (ID) data. While existing methods perform well on small-scale images, they suffer from mode collapse when synthesizing large-scale images, resulting in limited knowledge transfer. Recently, leveraging advanced generative models to synthesize photorealistic images has emerged as a promising alternative. Nevertheless, directly using off-the-shelf diffusion to generate datasets faces the precision-recall challenges: 1) ensuring synthetic data aligns with the real distribution, and 2) ensuring coverage of the real ID manifold. In response, we propose PRISM, a precision-recall informed synthesis method. Specifically, we introduce Energy-guided Distribution Alignment to avoid the generation of out-of-distribution samples, and design the Diversified Prompt Engineering to enhance coverage of the real ID manifold. Extensive experiments on various large-scale image datasets demonstrate the superiority of PRISM. Moreover, we demonstrate that models trained with PRISM exhibit strong domain generalization.
Similar Papers
PRISM: Diversifying Dataset Distillation by Decoupling Architectural Priors
Machine Learning (CS)
Creates better fake data for training computers.
PRISM: High-Resolution & Precise Counterfactual Medical Image Generation using Language-guided Stable Diffusion
CV and Pattern Recognition
Makes AI better at understanding medical pictures.
PRISM: Phase-enhanced Radial-based Image Signature Mapping framework for fingerprinting AI-generated images
CV and Pattern Recognition
Finds which AI made a picture.