Checklists Are Better Than Reward Models For Aligning Language Models
By: Vijay Viswanathan , Yanchao Sun , Shuang Ma and more
Potential Business Impact:
Teaches computers to follow all kinds of instructions.
Language models must be adapted to understand and follow user instructions. Reinforcement learning is widely used to facilitate this -- typically using fixed criteria such as "helpfulness" and "harmfulness". In our work, we instead propose using flexible, instruction-specific criteria as a means of broadening the impact that reinforcement learning can have in eliciting instruction following. We propose "Reinforcement Learning from Checklist Feedback" (RLCF). From instructions, we extract checklists and evaluate how well responses satisfy each item - using both AI judges and specialized verifier programs - then combine these scores to compute rewards for RL. We compare RLCF with other alignment methods applied to a strong instruction following model (Qwen2.5-7B-Instruct) on five widely-studied benchmarks -- RLCF is the only method to improve performance on every benchmark, including a 4-point boost in hard satisfaction rate on FollowBench, a 6-point increase on InFoBench, and a 3-point rise in win rate on Arena-Hard. These results establish checklist feedback as a key tool for improving language models' support of queries that express a multitude of needs.
Similar Papers
Instructions are all you need: Self-supervised Reinforcement Learning for Instruction Following
Computation and Language
Teaches computers to follow complex orders perfectly.
Rubric-Based Benchmarking and Reinforcement Learning for Advancing LLM Instruction Following
Computation and Language
Teaches AI to follow tricky, multi-step directions.
Explainable reinforcement learning from human feedback to improve alignment
Machine Learning (CS)
Fixes bad AI answers by finding and removing wrong training data.