GRAM: A Generative Foundation Reward Model for Reward Generalization
By: Chenglong Wang , Yang Gan , Yifu Huo and more
Potential Business Impact:
Teaches AI to learn better from more data.
In aligning large language models (LLMs), reward models have played an important role, but are standardly trained as discriminative models and rely only on labeled human preference data. In this paper, we explore methods that train reward models using both unlabeled and labeled data. Building on the generative models in LLMs, we develop a generative reward model that is first trained via large-scale unsupervised learning and then fine-tuned via supervised learning. We also show that by using label smoothing, we are in fact optimizing a regularized pairwise ranking loss. This result, in turn, provides a new view of training reward models, which links generative models and discriminative models under the same class of training objectives. The outcome of these techniques is a foundation reward model, which can be applied to a wide range of tasks with little or no further fine-tuning effort. Extensive experiments show that this model generalizes well across several tasks, including response ranking, reinforcement learning from human feedback, and task adaptation with fine-tuning, achieving significant performance improvements over several strong baseline models.
Similar Papers
GRAM-R$^2$: Self-Training Generative Foundation Reward Models for Reward Reasoning
Computation and Language
Teaches AI to explain why it likes answers.
GRAM-R$^2$: Self-Training Generative Foundation Reward Models for Reward Reasoning
Computation and Language
Teaches AI to explain why it picks answers.
Sentence-level Reward Model can Generalize Better for Aligning LLM from Human Preference
Computation and Language
Makes AI understand what people like better.