Beyond Consensus: Perspectivist Modeling and Evaluation of Annotator Disagreement in NLP
By: Yinuo Xu, David Jurgens
Annotator disagreement is widespread in NLP, particularly for subjective and ambiguous tasks such as toxicity detection and stance analysis. While early approaches treated disagreement as noise to be removed, recent work increasingly models it as a meaningful signal reflecting variation in interpretation and perspective. This survey provides a unified view of disagreement-aware NLP methods. We first present a domain-agnostic taxonomy of the sources of disagreement spanning data, task, and annotator factors. We then synthesize modeling approaches using a common framework defined by prediction targets and pooling structure, highlighting a shift from consensus learning toward explicitly modeling disagreement, and toward capturing structured relationships among annotators. We review evaluation metrics for both predictive performance and annotator behavior, and noting that most fairness evaluations remain descriptive rather than normative. We conclude by identifying open challenges and future directions, including integrating multiple sources of variation, developing disagreement-aware interpretability frameworks, and grappling with the practical tradeoffs of perspectivist modeling.
Similar Papers
Modeling Annotator Disagreement with Demographic-Aware Experts and Synthetic Perspectives
Computation and Language
Helps computers understand different people's opinions.
The Value of Disagreement in AI Design, Evaluation, and Alignment
Computers and Society
Makes AI fairer by including different ideas.
NUTMEG: Separating Signal From Noise in Annotator Disagreement
Computation and Language
Cleans up messy text data for better computer learning.