Parallel Latent Reasoning for Sequential Recommendation
By: Jiakai Tang , Xu Chen , Wen Chen and more
Potential Business Impact:
Finds what you like by trying many ideas.
Capturing complex user preferences from sparse behavioral sequences remains a fundamental challenge in sequential recommendation. Recent latent reasoning methods have shown promise by extending test-time computation through multi-step reasoning, yet they exclusively rely on depth-level scaling along a single trajectory, suffering from diminishing returns as reasoning depth increases. To address this limitation, we propose \textbf{Parallel Latent Reasoning (PLR)}, a novel framework that pioneers width-level computational scaling by exploring multiple diverse reasoning trajectories simultaneously. PLR constructs parallel reasoning streams through learnable trigger tokens in continuous latent space, preserves diversity across streams via global reasoning regularization, and adaptively synthesizes multi-stream outputs through mixture-of-reasoning-streams aggregation. Extensive experiments on three real-world datasets demonstrate that PLR substantially outperforms state-of-the-art baselines while maintaining real-time inference efficiency. Theoretical analysis further validates the effectiveness of parallel reasoning in improving generalization capability. Our work opens new avenues for enhancing reasoning capacity in sequential recommendation beyond existing depth scaling.
Similar Papers
LARES: Latent Reasoning for Sequential Recommendation
Information Retrieval
Helps computers guess what you'll like next.
A Survey on Parallel Reasoning
Computation and Language
Helps computers think in many ways at once.
Bridging Search and Recommendation through Latent Cross Reasoning
Information Retrieval
Finds better videos by understanding your searches.