DDTime: Dataset Distillation with Spectral Alignment and Information Bottleneck for Time-Series Forecasting
By: Yuqi Li , Kuiye Ding , Chuanguang Yang and more
Potential Business Impact:
Makes computer predictions faster with less data.
Time-series forecasting is fundamental across many domains, yet training accurate models often requires large-scale datasets and substantial computational resources. Dataset distillation offers a promising alternative by synthesizing compact datasets that preserve the learning behavior of full data. However, extending dataset distillation to time-series forecasting is non-trivial due to two fundamental challenges: 1.temporal bias from strong autocorrelation, which leads to distorted value-term alignment between teacher and student models; and 2.insufficient diversity among synthetic samples, arising from the absence of explicit categorical priors to regularize trajectory variety. In this work, we propose DDTime, a lightweight and plug-in distillation framework built upon first-order condensation decomposition. To tackle Challenge 1, it revisits value-term alignment through temporal statistics and introduces a frequency-domain alignment mechanism to mitigate autocorrelation-induced bias, ensuring spectral consistency and temporal fidelity. To address Challenge 2, we further design an inter-sample regularization inspired by the information bottleneck principle, which enhances diversity and maximizes information density across synthetic trajectories. The combined objective is theoretically compatible with a wide range of condensation paradigms and supports stable first-order optimization. Extensive experiments on 20 benchmark datasets and diverse forecasting architectures demonstrate that DDTime consistently outperforms existing distillation methods, achieving about 30% relative accuracy gains while introducing about 2.49% computational overhead. All code and distilled datasets will be released.
Similar Papers
Temporal Saliency-Guided Distillation: A Scalable Framework for Distilling Video Datasets
CV and Pattern Recognition
Makes big video files much smaller for computers.
TGDD: Trajectory Guided Dataset Distillation with Balanced Distribution
CV and Pattern Recognition
Makes computer learning faster and better.
TimeDistill: Efficient Long-Term Time Series Forecasting with MLP via Cross-Architecture Distillation
Machine Learning (CS)
Makes smart computer predictions faster and smaller.