Rotate Both Ways: Time-and-Order RoPE for Generative Recommendation
By: Xiaokai Wei , Jiajun Wu , Daiyao Yi and more
Potential Business Impact:
Helps apps guess what you'll do next.
Generative recommenders, typically transformer-based autoregressive models, predict the next item or action from a user's interaction history. Their effectiveness depends on how the model represents where an interaction event occurs in the sequence (discrete index) and when it occurred in wall-clock time. Prevailing approaches inject time via learned embeddings or relative attention biases. In this paper, we argue that RoPE-based approaches, if designed properly, can be a stronger alternative for jointly modeling temporal and sequential information in user behavior sequences. While vanilla RoPE in LLMs considers only token order, generative recommendation requires incorporating both event time and token index. To address this, we propose Time-and-Order RoPE (TO-RoPE), a family of rotary position embedding designs that treat index and time as angle sources shaping the query-key geometry directly. We present three instantiations: early fusion, split-by-dim, and split-by-head. Extensive experiments on both publicly available datasets and a proprietary industrial dataset show that TO-RoPE variants consistently improve accuracy over existing methods for encoding time and index. These results position rotary embeddings as a simple, principled, and deployment-friendly foundation for generative recommendation.
Similar Papers
Selective Rotary Position Embedding
Computation and Language
Makes AI better at remembering and understanding long stories.
HoPE: Hyperbolic Rotary Positional Encoding for Stable Long-Range Dependency Modeling in Large Language Models
Computation and Language
Makes AI understand long sentences better.
Rethinking RoPE: A Mathematical Blueprint for N-dimensional Positional Embedding
Machine Learning (CS)
Makes AI understand images and text better.