GRAM-R$^2$: Self-Training Generative Foundation Reward Models for Reward Reasoning
By: Chenglong Wang , Yongyu Mu , Hang Zhou and more
Potential Business Impact:
Teaches AI to explain why it likes answers.
Significant progress in reward modeling over recent years has been driven by a paradigm shift from task-specific designs towards generalist reward models. Despite this trend, developing effective reward models remains a fundamental challenge: the heavy reliance on large-scale labeled preference data. Pre-training on abundant unlabeled data offers a promising direction, but existing approaches fall short of instilling explicit reasoning into reward models. To bridge this gap, we propose a self-training approach that leverages unlabeled data to elicit reward reasoning in reward models. Based on this approach, we develop GRAM-R$^2$, a generative reward model trained to produce not only preference labels but also accompanying reward rationales. GRAM-R$^2$ can serve as a foundation model for reward reasoning and can be applied to a wide range of tasks with minimal or no additional fine-tuning. It can support downstream applications such as response ranking and task-specific reward tuning. Experiments on response ranking, task adaptation, and reinforcement learning from human feedback demonstrate that GRAM-R$^2$ consistently delivers strong performance, outperforming several strong discriminative and generative baselines.
Similar Papers
GRAM-R$^2$: Self-Training Generative Foundation Reward Models for Reward Reasoning
Computation and Language
Teaches AI to explain why it picks answers.
GRAM: A Generative Foundation Reward Model for Reward Generalization
Computation and Language
Teaches AI to learn better from more data.
MedGR$^2$: Breaking the Data Barrier for Medical Reasoning via Generative Reward Learning
Machine Learning (CS)
Makes AI learn medicine from generated data.