Paper Copilot: Tracking the Evolution of Peer Review in AI Conferences
By: Jing Yang, Qiyao Wei, Jiaxin Pei
Potential Business Impact:
Tracks how AI papers are judged fairly.
The rapid growth of AI conferences is straining an already fragile peer-review system, leading to heavy reviewer workloads, expertise mismatches, inconsistent evaluation standards, superficial or templated reviews, and limited accountability under compressed timelines. In response, conference organizers have introduced new policies and interventions to preserve review standards. Yet these ad-hoc changes often create further concerns and confusion about the review process, leaving how papers are ultimately accepted - and how practices evolve across years - largely opaque. We present Paper Copilot, a system that creates durable digital archives of peer reviews across a wide range of computer-science venues, an open dataset that enables researchers to study peer review at scale, and a large-scale empirical analysis of ICLR reviews spanning multiple years. By releasing both the infrastructure and the dataset, Paper Copilot supports reproducible research on the evolution of peer review. We hope these resources help the community track changes, diagnose failure modes, and inform evidence-based improvements toward a more robust, transparent, and reliable peer-review system.
Similar Papers
Position: The AI Conference Peer Review Crisis Demands Author Feedback and Reviewer Rewards
Artificial Intelligence
Makes AI paper reviews fairer and better.
Zero-shot reasoning for simulating scholarly peer-review
Artificial Intelligence
Checks science papers for AI cheating.
ReviewerToo: Should AI Join The Program Committee? A Look At The Future of Peer Review
Artificial Intelligence
AI helps scientists check research papers faster.