Score: 0

PRISM: Precision-Recall Informed Data-Free Knowledge Distillation via Generative Diffusion

Published: September 21, 2025 | arXiv ID: 2509.16897v1

By: Xuewan He , Jielei Wang , Zihan Cheng and more

Potential Business Impact:

Makes AI learn from fake pictures, not real ones.

Business Areas:
Visual Search Internet Services

Data-free knowledge distillation (DFKD) transfers knowledge from a teacher to a student without access to the real in-distribution (ID) data. While existing methods perform well on small-scale images, they suffer from mode collapse when synthesizing large-scale images, resulting in limited knowledge transfer. Recently, leveraging advanced generative models to synthesize photorealistic images has emerged as a promising alternative. Nevertheless, directly using off-the-shelf diffusion to generate datasets faces the precision-recall challenges: 1) ensuring synthetic data aligns with the real distribution, and 2) ensuring coverage of the real ID manifold. In response, we propose PRISM, a precision-recall informed synthesis method. Specifically, we introduce Energy-guided Distribution Alignment to avoid the generation of out-of-distribution samples, and design the Diversified Prompt Engineering to enhance coverage of the real ID manifold. Extensive experiments on various large-scale image datasets demonstrate the superiority of PRISM. Moreover, we demonstrate that models trained with PRISM exhibit strong domain generalization.

Page Count
5 pages

Category
Computer Science:
CV and Pattern Recognition