Score: 0

Explainable post-training bias mitigation with distribution-based fairness metrics

Published: April 1, 2025 | arXiv ID: 2504.01223v2

By: Ryan Franks, Alexey Miroshnikov, Konstandinos Kotsiopoulos

Potential Business Impact:

Makes AI fair without retraining.

Business Areas:
Predictive Analytics Artificial Intelligence, Data and Analytics, Software

We develop a novel optimization framework with distribution-based fairness constraints for efficiently producing demographically blind, explainable models across a wide range of fairness levels. This is accomplished through post-processing, avoiding the need for retraining. Our framework, which is based on stochastic gradient descent, can be applied to a wide range of model types, with a particular emphasis on the post-processing of gradient-boosted decision trees. Additionally, we design a broad class of interpretable global bias metrics compatible with our method by building on previous work. We empirically test our methodology on a variety of datasets and compare it to other methods.

Country of Origin
🇺🇸 United States

Page Count
45 pages

Category
Computer Science:
Machine Learning (CS)