Evaluating Bias in LLMs for Job-Resume Matching: Gender, Race, and Education
By: Hayate Iso , Pouya Pezeshkpour , Nikita Bhutani and more
Potential Business Impact:
AI hiring tools still favor certain schools.
Large Language Models (LLMs) offer the potential to automate hiring by matching job descriptions with candidate resumes, streamlining recruitment processes, and reducing operational costs. However, biases inherent in these models may lead to unfair hiring practices, reinforcing societal prejudices and undermining workplace diversity. This study examines the performance and fairness of LLMs in job-resume matching tasks within the English language and U.S. context. It evaluates how factors such as gender, race, and educational background influence model decisions, providing critical insights into the fairness and reliability of LLMs in HR applications. Our findings indicate that while recent models have reduced biases related to explicit attributes like gender and race, implicit biases concerning educational background remain significant. These results highlight the need for ongoing evaluation and the development of advanced bias mitigation strategies to ensure equitable hiring practices when using LLMs in industry settings.
Similar Papers
Where Should I Study? Biased Language Models Decide! Evaluating Fairness in LMs for Academic Recommendations
Computation and Language
AI unfairly favors rich countries and men.
Addressing Bias in LLMs: Strategies and Application to Fair AI-based Recruitment
Artificial Intelligence
Removes gender bias from hiring AI.
No LLM is Free From Bias: A Comprehensive Study of Bias Evaluation in Large Language Models
Computation and Language
Finds and fixes unfairness in AI language.