Repetitive Contrastive Learning Enhances Mamba's Selectivity in Time Series Prediction
By: Wenbo Yan, Hanzhong Cao, Ying Tan
Potential Business Impact:
Makes predictions better by focusing on important moments.
Long sequence prediction is a key challenge in time series forecasting. While Mamba-based models have shown strong performance due to their sequence selection capabilities, they still struggle with insufficient focus on critical time steps and incomplete noise suppression, caused by limited selective abilities. To address this, we introduce Repetitive Contrastive Learning (RCL), a token-level contrastive pretraining framework aimed at enhancing Mamba's selective capabilities. RCL pretrains a single Mamba block to strengthen its selective abilities and then transfers these pretrained parameters to initialize Mamba blocks in various backbone models, improving their temporal prediction performance. RCL uses sequence augmentation with Gaussian noise and applies inter-sequence and intra-sequence contrastive learning to help the Mamba module prioritize information-rich time steps while ignoring noisy ones. Extensive experiments show that RCL consistently boosts the performance of backbone models, surpassing existing methods and achieving state-of-the-art results. Additionally, we propose two metrics to quantify Mamba's selective capabilities, providing theoretical, qualitative, and quantitative evidence for the improvements brought by RCL.
Similar Papers
Mamaba Can Learn Low-Dimensional Targets In-Context via Test-Time Feature Learning
Machine Learning (CS)
Helps computers learn patterns from data faster.
Attention Mamba: Time Series Modeling with Adaptive Pooling Acceleration and Receptive Field Enhancements
Machine Learning (CS)
Predicts future events better by seeing more patterns.
Mamba Can Learn Low-Dimensional Targets In-Context via Test-Time Feature Learning
Machine Learning (CS)
Mamba learns patterns faster by finding key information.