Explainable post-training bias mitigation with distribution-based fairness metrics
By: Ryan Franks, Alexey Miroshnikov, Konstandinos Kotsiopoulos
Potential Business Impact:
Makes AI fair without retraining.
We develop a novel optimization framework with distribution-based fairness constraints for efficiently producing demographically blind, explainable models across a wide range of fairness levels. This is accomplished through post-processing, avoiding the need for retraining. Our framework, which is based on stochastic gradient descent, can be applied to a wide range of model types, with a particular emphasis on the post-processing of gradient-boosted decision trees. Additionally, we design a broad class of interpretable global bias metrics compatible with our method by building on previous work. We empirically test our methodology on a variety of datasets and compare it to other methods.
Similar Papers
General Post-Processing Framework for Fairness Adjustment of Machine Learning Models
Machine Learning (CS)
Makes computer decisions fair without retraining.
Explanations as Bias Detectors: A Critical Study of Local Post-hoc XAI Methods for Fairness Exploration
Artificial Intelligence
Finds unfairness in AI, making it more just.
Meta Optimality for Demographic Parity Constrained Regression via Post-Processing
Machine Learning (Stat)
Makes computer predictions fair for everyone.