Quantifying Query Fairness Under Unawareness
By: Thomas Jaenich , Alejandro Moreo , Alessandro Fabris and more
Potential Business Impact:
Makes search results fair for everyone.
Traditional ranking algorithms are designed to retrieve the most relevant items for a user's query, but they often inherit biases from data that can unfairly disadvantage vulnerable groups. Fairness in information access systems (IAS) is typically assessed by comparing the distribution of groups in a ranking to a target distribution, such as the overall group distribution in the dataset. These fairness metrics depend on knowing the true group labels for each item. However, when groups are defined by demographic or sensitive attributes, these labels are often unknown, leading to a setting known as "fairness under unawareness". To address this, group membership can be inferred using machine-learned classifiers, and group prevalence is estimated by counting the predicted labels. Unfortunately, such an estimation is known to be unreliable under dataset shift, compromising the accuracy of fairness evaluations. In this paper, we introduce a robust fairness estimator based on quantification that effectively handles multiple sensitive attributes beyond binary classifications. Our method outperforms existing baselines across various sensitive attributes and, to the best of our knowledge, is the first to establish a reliable protocol for measuring fairness under unawareness across multiple queries and groups.
Similar Papers
Fairness-aware Anomaly Detection via Fair Projection
Machine Learning (CS)
Makes AI fairer by spotting bad data for everyone.
Quantifying Group Fairness in Community Detection
Social and Information Networks
Finds unfairness in group networks, helps fix it.
Size-adaptive Hypothesis Testing for Fairness
Machine Learning (CS)
Checks if computer programs are unfair to groups.