Rating competitors in games with strength-dependent tie probabilities
By: Mark E. Glickman
Potential Business Impact:
Makes game ratings fairer by counting ties.
Competitor rating systems for head-to-head games are typically used to measure playing strength from game outcomes. Ratings computed from these systems are often used to select top competitors for elite events, for pairing players of similar strength in online gaming, and for players to track their own strength over time. Most implemented rating systems assume only win/loss outcomes, and treat occurrences of ties as the equivalent to half a win and half a loss. However, in games such as chess, the probability of a tie (draw) is demonstrably higher for stronger players than for weaker players, so that rating systems ignoring this aspect of game results may produce strength estimates that are unreliable. We develop a new rating system for head-to-head games based on a model by Glickman (2025) that explicitly acknowledges that a tie may depend on the strengths of the competitors. The approach uses a Bayesian dynamic modeling framework. Within each time period, posterior updates are computed in closed form using a single Newton-Raphson iteration evaluated at the prior mean. The approach is demonstrated on a large dataset of chess games played in International Correspondence Chess Federation tournaments.
Similar Papers
Drawing Conclusions from Draws: Rethinking Preference Semantics in Arena-Style LLM Evaluation
Computation and Language
Makes AI rating systems fairer by understanding "draws."
Tied Pools and Drawn Games
Statistics Theory
Figures out who's best when games can be draws.
Model inference for ranking from pairwise comparisons
Social and Information Networks
Figures out who's best from messy game results.