Significativity Indices for Agreement Values
By: Alberto Casagrande , Francesco Fabris , Rossano Girometti and more
Potential Business Impact:
Helps check if different computer programs agree.
Agreement measures, such as Cohen's kappa or intraclass correlation, gauge the matching between two or more classifiers. They are used in a wide range of contexts from medicine, where they evaluate the effectiveness of medical treatments and clinical trials, to artificial intelligence, where they can quantify the approximation due to the reduction of a classifier. The consistency of different classifiers to a golden standard can be compared simply by using the order induced by their agreement measure with respect to the golden standard itself. Nevertheless, labelling an approach as good or bad exclusively by using the value of an agreement measure requires a scale or a significativity index. Some quality scales have been proposed in the literature for Cohen's kappa, but they are mainly na\"ive, and their boundaries are arbitrary. This work proposes a general approach to evaluate the significativity of any agreement value between two classifiers and introduces two significativity indices: one dealing with finite data sets, the other one handling classification probability distributions. Moreover, this manuscript addresses the computational challenges of evaluating such indices and proposes some efficient algorithms for their evaluation.
Similar Papers
Fiducial Confidence Intervals for Agreement Measures Among Raters Under a Generalized Linear Mixed Effects Model
Methodology
Measures how well different people agree.
Scale-Insensitive Neural Network Significance Tests
Machine Learning (Stat)
Makes AI learn better without needing perfect data.
Absolute indices for determining compactness, separability and number of clusters
Machine Learning (CS)
Finds the best groups in data.