Modeling LLM Agent Reviewer Dynamics in Elo-Ranked Review System
By: Hsiang-Wei Huang , Junbin Lu , Kuang-Ming Chen and more
In this work, we explore the Large Language Model (LLM) agent reviewer dynamics in an Elo-ranked review system using real-world conference paper submissions. Multiple LLM agent reviewers with different personas are engage in multi round review interactions moderated by an Area Chair. We compare a baseline setting with conditions that incorporate Elo ratings and reviewer memory. Our simulation results showcase several interesting findings, including how incorporating Elo improves Area Chair decision accuracy, as well as reviewers' adaptive review strategy that exploits our Elo system without improving review effort. Our code is available at https://github.com/hsiangwei0903/EloReview.
Similar Papers
Scoring, Reasoning, and Selecting the Best! Ensembling Large Language Models via a Peer-Review Process
Computation and Language
Chooses the best AI answer from many.
LLM-REVal: Can We Trust LLM Reviewers Yet?
Computation and Language
AI reviewers unfairly favor AI-written papers.
Towards Simulating Social Influence Dynamics with LLM-based Multi-agents
Multiagent Systems
Computers can now act like people talking online.