Structured Reasoning for Fairness: A Multi-Agent Approach to Bias Detection in Textual Data
By: Tianyi Huang, Elsa Fan
Potential Business Impact:
Finds and fixes unfairness in AI writing.
From disinformation spread by AI chatbots to AI recommendations that inadvertently reinforce stereotypes, textual bias poses a significant challenge to the trustworthiness of large language models (LLMs). In this paper, we propose a multi-agent framework that systematically identifies biases by disentangling each statement as fact or opinion, assigning a bias intensity score, and providing concise, factual justifications. Evaluated on 1,500 samples from the WikiNPOV dataset, the framework achieves 84.9% accuracy$\unicode{x2014}$an improvement of 13.0% over the zero-shot baseline$\unicode{x2014}$demonstrating the efficacy of explicitly modeling fact versus opinion prior to quantifying bias intensity. By combining enhanced detection accuracy with interpretable explanations, this approach sets a foundation for promoting fairness and accountability in modern language models.
Similar Papers
BIASINSPECTOR: Detecting Bias in Structured Data through LLM Agents
Artificial Intelligence
Finds unfairness in computer data automatically.
FairReason: Balancing Reasoning and Social Bias in MLLMs
Artificial Intelligence
Makes AI smarter without being unfair.
Bias Mitigation Agent: Optimizing Source Selection for Fair and Balanced Knowledge Retrieval
Artificial Intelligence
Cleans AI answers to be fair and true.