UC-MOA: Utility-Conditioned Multi-Objective Alignment for Distributional Pareto-Optimality
By: Zelei Cheng , Xin-Qiang Cai , Yuting Tang and more
Potential Business Impact:
Teaches AI to better understand what people want.
Reinforcement Learning from Human Feedback (RLHF) has become a cornerstone for aligning large language models (LLMs) with human values. However, existing approaches struggle to capture the multi-dimensional, distributional nuances of human preferences. Methods such as RiC that directly inject raw reward values into prompts face significant numerical sensitivity issues--for instance, LLMs may fail to distinguish between 9.11 and 9.8--while alternatives like MORLHF, Rewarded Soups, and MODPO incur high computational costs by training multiple models. In this work, we introduce Utility-Conditioned Multi-Objective Alignment (UC-MOA), a novel framework that overcomes these limitations. Our approach leverages a diverse set of strictly increasing, non-linear utility functions to transform user-specified preferences into symbolic tokens, which are then used to condition a single LLM. This design not only mitigates numerical reasoning challenges but also substantially reduces training overhead, yielding models that achieve superior Pareto fronts and robust alignment across complex reward dimensions.
Similar Papers
Pareto Multi-Objective Alignment for Language Models
Machine Learning (CS)
Helps AI learn to balance many different goals.
MOA: Multi-Objective Alignment for Role-Playing Agents
Computation and Language
Teaches AI to be good at many things at once.
Multi-Objective Reinforcement Learning for Large Language Model Optimization: Visionary Perspective
Computation and Language
Teaches AI to do many things well.