Scalable Sequential Recommendation under Latency and Memory Constraints
By: Adithya Parthasarathy , Aswathnarayan Muthukrishnan Kirubakaran , Vinoth Punniyamoorthy and more
Sequential recommender systems must model long-range user behavior while operating under strict memory and latency constraints. Transformer-based approaches achieve strong accuracy but suffer from quadratic attention complexity, forcing aggressive truncation of user histories and limiting their practicality for long-horizon modeling. This paper presents HoloMambaRec, a lightweight sequential recommendation architecture that combines holographic reduced representations for attribute-aware embedding with a selective state space encoder for linear-time sequence processing. Item and attribute information are bound using circular convolution, preserving embedding dimensionality while encoding structured metadata. A shallow selective state space backbone, inspired by recent Mamba-style models, enables efficient training and constant-time recurrent inference. Experiments on Amazon Beauty and MovieLens-1M datasets demonstrate that HoloMambaRec consistently outperforms SASRec and achieves competitive performance with GRU4Rec under a constrained 10-epoch training budget, while maintaining substantially lower memory complexity. The design further incorporates forward-compatible mechanisms for temporal bundling and inference-time compression, positioning HoloMambaRec as a practical and extensible alternative for scalable, metadata-aware sequential recommendation.
Similar Papers
HMamba: Hyperbolic Mamba for Sequential Recommendation
Information Retrieval
Helps websites show you what you'll like next.
A Novel Mamba-based Sequential Recommendation Method
Information Retrieval
Recommends items faster with less computer power.
MMA: A Momentum Mamba Architecture for Human Activity Recognition with Inertial Sensors
Human-Computer Interaction
Helps computers understand body movements better.