Score: 1

EDITS: Enhancing Dataset Distillation with Implicit Textual Semantics

Published: September 17, 2025 | arXiv ID: 2509.13858v1

By: Qianxin Xia , Jiawei Du , Guoming Lu and more

Potential Business Impact:

Makes small data learn like big data.

Business Areas:
Text Analytics Data and Analytics, Software

Dataset distillation aims to synthesize a compact dataset from the original large-scale one, enabling highly efficient learning while preserving competitive model performance. However, traditional techniques primarily capture low-level visual features, neglecting the high-level semantic and structural information inherent in images. In this paper, we propose EDITS, a novel framework that exploits the implicit textual semantics within the image data to achieve enhanced distillation. First, external texts generated by a Vision Language Model (VLM) are fused with image features through a Global Semantic Query module, forming the prior clustered buffer. Local Semantic Awareness then selects representative samples from the buffer to construct image and text prototypes, with the latter produced by guiding a Large Language Model (LLM) with meticulously crafted prompt. Ultimately, Dual Prototype Guidance strategy generates the final synthetic dataset through a diffusion model. Extensive experiments confirm the effectiveness of our method.Source code is available in: https://github.com/einsteinxia/EDITS.

Repos / Data Links

Page Count
5 pages

Category
Computer Science:
CV and Pattern Recognition