FAIR: Focused Attention Is All You Need for Generative Recommendation
By: Longtao Xiao , Haolin Zhang , Guohao Cai and more
Recently, transformer-based generative recommendation has garnered significant attention for user behavior modeling. However, it often requires discretizing items into multi-code representations (e.g., typically four code tokens or more), which sharply increases the length of the original item sequence. This expansion poses challenges to transformer-based models for modeling user behavior sequences with inherent noises, since they tend to overallocate attention to irrelevant or noisy context. To mitigate this issue, we propose FAIR, the first generative recommendation framework with focused attention, which enhances attention scores to relevant context while suppressing those to irrelevant ones. Specifically, we propose (1) a focused attention mechanism integrated into the standard Transformer, which learns two separate sets of Q and K attention weights and computes their difference as the final attention scores to eliminate attention noise while focusing on relevant contexts; (2) a noise-robustness objective, which encourages the model to maintain stable attention patterns under stochastic perturbations, preventing undesirable shifts toward irrelevant context due to noise; and (3) a mutual information maximization objective, which guides the model to identify contexts that are most informative for next-item prediction. We validate the effectiveness of FAIR on four public benchmarks, demonstrating its superior performance compared to existing methods.
Similar Papers
Why Generate When You Can Transform? Unleashing Generative Attention for Dynamic Recommendation
Information Retrieval
Helps apps guess what you'll like next.
Learning to Focus: Focal Attention for Selective and Scalable Transformers
Computation and Language
Makes AI focus better on important words.
From Scaling to Structured Expressivity: Rethinking Transformers for CTR Prediction
Information Retrieval
Helps online ads show better by understanding user choices.