A Gray Literature Study on Fairness Requirements in AI-enabled Software Engineering
By: Thanh Nguyen, Chaima Boufaied, Ronnie de Souza Santos
Potential Business Impact:
Makes AI fair, not just smart.
Today, with the growing obsession with applying Artificial Intelligence (AI), particularly Machine Learning (ML), to software across various contexts, much of the focus has been on the effectiveness of AI models, often measured through common metrics such as F1- score, while fairness receives relatively little attention. This paper presents a review of existing gray literature, examining fairness requirements in AI context, with a focus on how they are defined across various application domains, managed throughout the Software Development Life Cycle (SDLC), and the causes, as well as the corresponding consequences of their violation by AI models. Our gray literature investigation shows various definitions of fairness requirements in AI systems, commonly emphasizing non-discrimination and equal treatment across different demographic and social attributes. Fairness requirement management practices vary across the SDLC, particularly in model training and bias mitigation, fairness monitoring and evaluation, and data handling practices. Fairness requirement violations are frequently linked, but not limited, to data representation bias, algorithmic and model design bias, human judgment, and evaluation and transparency gaps. The corresponding consequences include harm in a broad sense, encompassing specific professional and societal impacts as key examples, stereotype reinforcement, data and privacy risks, and loss of trust and legitimacy in AI-supported decisions. These findings emphasize the need for consistent frameworks and practices to integrate fairness into AI software, paying as much attention to fairness as to effectiveness.
Similar Papers
Practitioner Insights on Fairness Requirements in the AI Development Life Cycle: An Interview Study
Software Engineering
Helps make AI fair for everyone.
On the Robustness of Fairness Practices: A Causal Framework for Systematic Evaluation
Software Engineering
Makes computer decisions fair for everyone.
Exploring the Landscape of Fairness Interventions in Software Engineering
Software Engineering
Makes AI fair, avoiding bad decisions.