Evaluating Judges as Evaluators: The JETTS Benchmark of LLM-as-Judges as Test-Time Scaling Evaluators
By: Yilun Zhou , Austin Xu , Peifeng Wang and more
Potential Business Impact:
Helps AI learn to fix its own mistakes.
Scaling test-time computation, or affording a generator large language model (LLM) extra compute during inference, typically employs the help of external non-generative evaluators (i.e., reward models). Concurrently, LLM-judges, models trained to generate evaluations and critiques (explanations) in natural language, are becoming increasingly popular in automatic evaluation. Despite judge empirical successes, their effectiveness as evaluators in test-time scaling settings is largely unknown. In this paper, we introduce the Judge Evaluation for Test-Time Scaling (JETTS) benchmark, which evaluates judge performance in three domains (math reasoning, code generation, and instruction following) under three task settings: response reranking, step-level beam search, and critique-based response refinement. We evaluate 10 different judge models (7B-70B parameters) for 8 different base generator models (6.7B-72B parameters). Our benchmark shows that while judges are competitive with outcome reward models in reranking, they are consistently worse than process reward models in beam search procedures. Furthermore, though unique to LLM-judges, their natural language critiques are currently ineffective in guiding the generator towards better responses.
Similar Papers
J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge
Machine Learning (CS)
Helps AI explain its answers better.
An Empirical Study of LLM-as-a-Judge: How Design Choices Impact Evaluation Reliability
Computation and Language
Helps computers judge other computers' answers.
Are We on the Right Way to Assessing LLM-as-a-Judge?
Computation and Language
Checks if AI judges are fair and honest.