Data-free Knowledge Distillation with Diffusion Models
By: Xiaohua Qi , Renda Li , Long Peng and more
Potential Business Impact:
Teaches computers new skills without needing old examples.
Recently Data-Free Knowledge Distillation (DFKD) has garnered attention and can transfer knowledge from a teacher neural network to a student neural network without requiring any access to training data. Although diffusion models are adept at synthesizing high-fidelity photorealistic images across various domains, existing methods cannot be easiliy implemented to DFKD. To bridge that gap, this paper proposes a novel approach based on diffusion models, DiffDFKD. Specifically, DiffDFKD involves targeted optimizations in two key areas. Firstly, DiffDFKD utilizes valuable information from teacher models to guide the pre-trained diffusion models' data synthesis, generating datasets that mirror the training data distribution and effectively bridge domain gaps. Secondly, to reduce computational burdens, DiffDFKD introduces Latent CutMix Augmentation, an efficient technique, to enhance the diversity of diffusion model-generated images for DFKD while preserving key attributes for effective knowledge transfer. Extensive experiments validate the efficacy of DiffDFKD, yielding state-of-the-art results exceeding existing DFKD approaches. We release our code at https://github.com/xhqi0109/DiffDFKD.
Similar Papers
CAE-DFKD: Bridging the Transferability Gap in Data-Free Knowledge Distillation
CV and Pattern Recognition
Teaches computers new skills without needing old examples.
Adversarial Curriculum Graph-Free Knowledge Distillation for Graph Neural Networks
Machine Learning (CS)
Teaches computers to learn without needing real examples.
A Comprehensive Survey on Knowledge Distillation
CV and Pattern Recognition
Makes big AI models run on small devices.