EDITS: Enhancing Dataset Distillation with Implicit Textual Semantics
By: Qianxin Xia , Jiawei Du , Guoming Lu and more
Potential Business Impact:
Makes small data learn like big data.
Dataset distillation aims to synthesize a compact dataset from the original large-scale one, enabling highly efficient learning while preserving competitive model performance. However, traditional techniques primarily capture low-level visual features, neglecting the high-level semantic and structural information inherent in images. In this paper, we propose EDITS, a novel framework that exploits the implicit textual semantics within the image data to achieve enhanced distillation. First, external texts generated by a Vision Language Model (VLM) are fused with image features through a Global Semantic Query module, forming the prior clustered buffer. Local Semantic Awareness then selects representative samples from the buffer to construct image and text prototypes, with the latter produced by guiding a Large Language Model (LLM) with meticulously crafted prompt. Ultimately, Dual Prototype Guidance strategy generates the final synthetic dataset through a diffusion model. Extensive experiments confirm the effectiveness of our method.Source code is available in: https://github.com/einsteinxia/EDITS.
Similar Papers
Technical Report on Text Dataset Distillation
Machine Learning (CS)
Creates small text sets that teach computers well.
Leveraging Multi-Modal Information to Enhance Dataset Distillation
CV and Pattern Recognition
Makes fake pictures teach computers better.
Dataset Distillation for Pre-Trained Self-Supervised Vision Models
CV and Pattern Recognition
Creates small, smart picture sets for AI.