Score: 1

MGD$^3$: Mode-Guided Dataset Distillation using Diffusion Models

Published: May 25, 2025 | arXiv ID: 2505.18963v1

By: Jeffrey A. Chan-Santiago , Praveen Tirupattur , Gaurav Kumar Nayak and more

Potential Business Impact:

Makes AI learn faster with better fake data.

Business Areas:
Predictive Analytics Artificial Intelligence, Data and Analytics, Software

Dataset distillation has emerged as an effective strategy, significantly reducing training costs and facilitating more efficient model deployment. Recent advances have leveraged generative models to distill datasets by capturing the underlying data distribution. Unfortunately, existing methods require model fine-tuning with distillation losses to encourage diversity and representativeness. However, these methods do not guarantee sample diversity, limiting their performance. We propose a mode-guided diffusion model leveraging a pre-trained diffusion model without the need to fine-tune with distillation losses. Our approach addresses dataset diversity in three stages: Mode Discovery to identify distinct data modes, Mode Guidance to enhance intra-class diversity, and Stop Guidance to mitigate artifacts in synthetic samples that affect performance. Our approach outperforms state-of-the-art methods, achieving accuracy gains of 4.4%, 2.9%, 1.6%, and 1.6% on ImageNette, ImageIDC, ImageNet-100, and ImageNet-1K, respectively. Our method eliminates the need for fine-tuning diffusion models with distillation losses, significantly reducing computational costs. Our code is available on the project webpage: https://jachansantiago.github.io/mode-guided-distillation/

Country of Origin
πŸ‡ΊπŸ‡Έ United States

Page Count
16 pages

Category
Computer Science:
CV and Pattern Recognition