Algorithmic Accountability in Small Data: Sample-Size-Induced Bias Within Classification Metrics
By: Jarren Briscoe , Garrett Kepler , Daryl Deford and more
Potential Business Impact:
Fixes unfair computer decisions when groups are different sizes.
Evaluating machine learning models is crucial not only for determining their technical accuracy but also for assessing their potential societal implications. While the potential for low-sample-size bias in algorithms is well known, we demonstrate the significance of sample-size bias induced by combinatorics in classification metrics. This revelation challenges the efficacy of these metrics in assessing bias with high resolution, especially when comparing groups of disparate sizes, which frequently arise in social applications. We provide analyses of the bias that appears in several commonly applied metrics and propose a model-agnostic assessment and correction technique. Additionally, we analyze counts of undefined cases in metric calculations, which can lead to misleading evaluations if improperly handled. This work illuminates the previously unrecognized challenge of combinatorics and probability in standard evaluation practices and thereby advances approaches for performing fair and trustworthy classification methods.
Similar Papers
A Survey on Small Sample Imbalance Problem: Metrics, Feature Analysis, and Solutions
Machine Learning (CS)
Helps computers learn better from tricky, unbalanced data.
On the Origins of Sampling Bias: Implications on Fairness Measurement and Mitigation
Machine Learning (CS)
Fixes unfairness in computer learning.
Measures of classification bias derived from sample size analysis
Methodology
Finds unfair computer decisions faster.