Score: 1

FaCTR: Factorized Channel-Temporal Representation Transformers for Efficient Time Series Forecasting

Published: June 5, 2025 | arXiv ID: 2506.05597v1

By: Yash Vijay, Harini Subramanyan

BigTech Affiliations: C3.ai

Potential Business Impact:

Predicts future events more accurately and simply.

Business Areas:
Facial Recognition Data and Analytics, Software

While Transformers excel in language and vision-where inputs are semantically rich and exhibit univariate dependency structures-their architectural complexity leads to diminishing returns in time series forecasting. Time series data is characterized by low per-timestep information density and complex dependencies across channels and covariates, requiring conditioning on structured variable interactions. To address this mismatch and overparameterization, we propose FaCTR, a lightweight spatiotemporal Transformer with an explicitly structural design. FaCTR injects dynamic, symmetric cross-channel interactions-modeled via a low-rank Factorization Machine into temporally contextualized patch embeddings through a learnable gating mechanism. It further encodes static and dynamic covariates for multivariate conditioning. Despite its compact design, FaCTR achieves state-of-the-art performance on eleven public forecasting benchmarks spanning both short-term and long-term horizons, with its largest variant using close to only 400K parameters-on average 50x smaller than competitive spatiotemporal transformer baselines. In addition, its structured design enables interpretability through cross-channel influence scores-an essential requirement for real-world decision-making. Finally, FaCTR supports self-supervised pretraining, positioning it as a compact yet versatile foundation for downstream time series tasks.

Country of Origin
πŸ‡ΊπŸ‡Έ United States

Page Count
27 pages

Category
Computer Science:
Machine Learning (CS)