From 1,000,000 Users to Every User: Scaling Up Personalized Preference for User-level Alignment
By: Jia-Nan Li , Jian Guan , Songhao Wu and more
Potential Business Impact:
Teaches AI to understand what *you* want.
Large language models (LLMs) have traditionally been aligned through one-size-fits-all approaches that assume uniform human preferences, fundamentally overlooking the diversity in user values and needs. This paper introduces a comprehensive framework for scalable personalized alignment of LLMs. We establish a systematic preference space characterizing psychological and behavioral dimensions, alongside diverse persona representations for robust preference inference in real-world scenarios. Building upon this foundation, we introduce \textsc{AlignX}, a large-scale dataset of over 1.3 million personalized preference examples, and develop two complementary alignment approaches: \textit{in-context alignment} directly conditioning on persona representations and \textit{preference-bridged alignment} modeling intermediate preference distributions. Extensive experiments demonstrate substantial improvements over existing methods, with an average 17.06\% accuracy gain across four benchmarks while exhibiting a strong adaptation capability to novel preferences, robustness to limited user data, and precise preference controllability. These results validate our approach toward user-adaptive AI systems.
Similar Papers
A Survey on Personalized Alignment -- The Missing Piece for Large Language Models in Real-World Applications
Computation and Language
Teaches AI to be helpful and kind, your way.
A Survey on Personalized and Pluralistic Preference Alignment in Large Language Models
Computation and Language
Makes AI understand what you like best.
HyPerAlign: Interpretable Personalized LLM Alignment via Hypothesis Generation
Computation and Language
Makes AI write like you, not everyone.