L-GTA: Latent Generative Modeling for Time Series Augmentation
By: Luis Roque , Carlos Soares , Vitor Cerqueira and more
Potential Business Impact:
Makes computer predictions better by creating fake data.
Data augmentation is gaining importance across various aspects of time series analysis, from forecasting to classification and anomaly detection tasks. We introduce the Latent Generative Transformer Augmentation (L-GTA) model, a generative approach using a transformer-based variational recurrent autoencoder. This model uses controlled transformations within the latent space of the model to generate new time series that preserve the intrinsic properties of the original dataset. L-GTA enables the application of diverse transformations, ranging from simple jittering to magnitude warping, and combining these basic transformations to generate more complex synthetic time series datasets. Our evaluation of several real-world datasets demonstrates the ability of L-GTA to produce more reliable, consistent, and controllable augmented data. This translates into significant improvements in predictive accuracy and similarity measures compared to direct transformation methods.
Similar Papers
Self-Supervised Temporal Super-Resolution of Energy Data using Generative Adversarial Transformer
Machine Learning (CS)
Makes energy grids work better with less data.
Self-Supervised Temporal Super-Resolution of Energy Data using Generative Adversarial Transformer
Machine Learning (CS)
Makes energy grids smarter and more accurate.
Less Is More: Generating Time Series with LLaMA-Style Autoregression in Simple Factorized Latent Spaces
Machine Learning (CS)
Makes fake data that looks real, fast.