EVALUESTEER: Measuring Reward Model Steerability Towards Values and Preference
By: Kshitish Ghate , Andy Liu , Devansh Jain and more
Potential Business Impact:
Helps AI understand what you like and want.
As large language models (LLMs) are deployed globally, creating pluralistic systems that can accommodate the diverse preferences and values of users worldwide becomes essential. We introduce EVALUESTEER, a benchmark to measure LLMs' and reward models' (RMs) steerability towards users' value and stylistic preference profiles grounded in psychology and human-LLM interaction literature. To address the gap in existing datasets that do not support controlled evaluations of RM steering, we synthetically generated 165,888 preference pairs -- systematically varying pairs along 4 value dimensions (traditional, secular-rational, survival, and self-expression) and 4 style dimensions (verbosity, readability, confidence, and warmth). We use EVALUESTEER to evaluate whether, given a user profile and a pair of candidate value-laden and style-laden responses, LLMs and RMs are able to select the output that aligns with the user's preferences. We evaluate six open-source and proprietary LLMs and RMs under sixteen systematic prompting conditions and six preference comparison scenarios. Notably, our results show that, when given the user's full profile of values and stylistic preferences, the best models achieve <75% accuracy at choosing the correct response, in contrast to >99% accuracy when only relevant style and value preferences are provided. EVALUESTEER thus highlights the limitations of current RMs at identifying and adapting to relevant user profile information, and provides a challenging testbed for developing RMs that can be steered towards diverse human values and preferences.
Similar Papers
STEER-BENCH: A Benchmark for Evaluating the Steerability of Large Language Models
Computation and Language
Helps computers understand different online groups.
STEER-ME: Assessing the Microeconomic Reasoning of Large Language Models
Computation and Language
Tests if computers understand money and business.
A Course Correction in Steerability Evaluation: Revealing Miscalibration and Side Effects in LLMs
Computation and Language
Helps computers follow all your instructions perfectly.