FAIRE: Assessing Racial and Gender Bias in AI-Driven Resume Evaluations
By: Athena Wen , Tanush Patil , Ansh Saxena and more
Potential Business Impact:
Tests AI hiring tools for race and gender bias.
In an era where AI-driven hiring is transforming recruitment practices, concerns about fairness and bias have become increasingly important. To explore these issues, we introduce a benchmark, FAIRE (Fairness Assessment In Resume Evaluation), to test for racial and gender bias in large language models (LLMs) used to evaluate resumes across different industries. We use two methods-direct scoring and ranking-to measure how model performance changes when resumes are slightly altered to reflect different racial or gender identities. Our findings reveal that while every model exhibits some degree of bias, the magnitude and direction vary considerably. This benchmark provides a clear way to examine these differences and offers valuable insights into the fairness of AI-based hiring tools. It highlights the urgent need for strategies to reduce bias in AI-driven recruitment. Our benchmark code and dataset are open-sourced at our repository: https://github.com/athenawen/FAIRE-Fairness-Assessment-In-Resume-Evaluation.git.
Similar Papers
Fairness Is Not Enough: Auditing Competence and Intersectional Bias in AI-powered Resume Screening
Computers and Society
AI hiring tools can be unfair or bad at jobs.
Where Fact Ends and Fairness Begins: Redefining AI Bias Evaluation through Cognitive Biases
Computation and Language
Tests AI for truth and fairness.
Evaluating Bias in LLMs for Job-Resume Matching: Gender, Race, and Education
Computation and Language
AI hiring tools still favor certain schools.