Epistemology gives a Future to Complementarity in Human-AI Interactions
By: Andrea Ferrario, Alessandro Facchini, Juan M. Durán
Human-AI complementarity is the claim that a human supported by an AI system can outperform either alone in a decision-making process. Since its introduction in the human-AI interaction literature, it has gained traction by generalizing the reliance paradigm and by offering a more practical alternative to the contested construct of 'trust in AI.' Yet complementarity faces key theoretical challenges: it lacks precise theoretical anchoring, it is formalized just as a post hoc indicator of relative predictive accuracy, it remains silent about other desiderata of human-AI interactions and it abstracts away from the magnitude-cost profile of its performance gain. As a result, complementarity is difficult to obtain in empirical settings. In this work, we leverage epistemology to address these challenges by reframing complementarity within the discourse on justificatory AI. Drawing on computational reliabilism, we argue that historical instances of complementarity function as evidence that a given human-AI interaction is a reliable epistemic process for a given predictive task. Together with other reliability indicators assessing the alignment of the human-AI team with the epistemic standards and socio-technical practices, complementarity contributes to the degree of reliability of human-AI teams when generating predictions. This supports the practical reasoning of those affected by these outputs -- patients, managers, regulators, and others. In summary, our approach suggests that the role and value of complementarity lies not in providing a relative measure of predictive accuracy, but in helping calibrate decision-making to the reliability of AI-supported processes that increasingly shape everyday life.
Similar Papers
Classifying Epistemic Relationships in Human-AI Interaction: An Exploratory Approach
Human-Computer Interaction
Helps people work better with smart computer helpers.
Human-AI Complementarity: A Goal for Amplified Oversight
Artificial Intelligence
AI helps people check if AI is telling the truth.
Fostering human learning is crucial for boosting human-AI synergy
Human-Computer Interaction
Helps people and computers work better together.