Revisiting Pre-processing Group Fairness: A Modular Benchmarking Framework
By: Brodie Oldfield, Ziqi Xu, Sevvandi Kandanaarachchi
Potential Business Impact:
Makes computer decisions fairer by fixing bad data.
As machine learning systems become increasingly integrated into high-stakes decision-making processes, ensuring fairness in algorithmic outcomes has become a critical concern. Methods to mitigate bias typically fall into three categories: pre-processing, in-processing, and post-processing. While significant attention has been devoted to the latter two, pre-processing methods, which operate at the data level and offer advantages such as model-agnosticism and improved privacy compliance, have received comparatively less focus and lack standardised evaluation tools. In this work, we introduce FairPrep, an extensible and modular benchmarking framework designed to evaluate fairness-aware pre-processing techniques on tabular datasets. Built on the AIF360 platform, FairPrep allows seamless integration of datasets, fairness interventions, and predictive models. It features a batch-processing interface that enables efficient experimentation and automatic reporting of fairness and utility metrics. By offering standardised pipelines and supporting reproducible evaluations, FairPrep fills a critical gap in the fairness benchmarking landscape and provides a practical foundation for advancing data-level fairness research.
Similar Papers
General Post-Processing Framework for Fairness Adjustment of Machine Learning Models
Machine Learning (CS)
Makes computer decisions fair without retraining.
CausalPre: Scalable and Effective Data Pre-processing for Causal Fairness
Machine Learning (CS)
Makes computer decisions fairer by understanding how things connect.
Software Fairness Dilemma: Is Bias Mitigation a Zero-Sum Game?
Machine Learning (CS)
Makes AI fairer without hurting anyone's performance.