PLUM: Adapting Pre-trained Language Models for Industrial-scale Generative Recommendations
By: Ruining He , Lukasz Heldt , Lichan Hong and more
Potential Business Impact:
Helps YouTube suggest videos you'll love.
Large Language Models (LLMs) pose a new paradigm of modeling and computation for information tasks. Recommendation systems are a critical application domain poised to benefit significantly from the sequence modeling capabilities and world knowledge inherent in these large models. In this paper, we introduce PLUM, a framework designed to adapt pre-trained LLMs for industry-scale recommendation tasks. PLUM consists of item tokenization using Semantic IDs, continued pre-training (CPT) on domain-specific data, and task-specific fine-tuning for recommendation objectives. For fine-tuning, we focus particularly on generative retrieval, where the model is directly trained to generate Semantic IDs of recommended items based on user context. We conduct comprehensive experiments on large-scale internal video recommendation datasets. Our results demonstrate that PLUM achieves substantial improvements for retrieval compared to a heavily-optimized production model built with large embedding tables. We also present a scaling study for the model's retrieval performance, our learnings about CPT, a few enhancements to Semantic IDs, along with an overview of the training and inference methods that enable launching this framework to billions of users in YouTube.
Similar Papers
Architecture is All You Need: Improving LLM Recommenders by Dropping the Text
Information Retrieval
Makes movie suggestions better with less computer power.
End-to-End Personalization: Unifying Recommender Systems with Large Language Models
Information Retrieval
Suggests movies you'll love, explains why.
Enhancing Recommender Systems Using Textual Embeddings from Pre-trained Language Models
Information Retrieval
Makes movie suggestions understand what you like.