A Profit-Based Measure of Lending Discrimination
By: Madison Coots , Robert Bartlett , Julian Nyarko and more
Algorithmic lending has transformed the consumer credit landscape, with complex machine learning models now commonly used to make or assist underwriting decisions. To comply with fair lending laws, these algorithms typically exclude legally protected characteristics, such as race and gender. Yet algorithmic underwriting can still inadvertently favor certain groups, prompting new questions about how to audit lending algorithms for potentially discriminatory behavior. Building on prior theoretical work, we introduce a profit-based measure of lending discrimination in loan pricing. Applying our approach to approximately 80,000 personal loans from a major U.S. fintech platform, we find that loans made to men and Black borrowers yielded lower profits than loans to other groups, indicating that men and Black applicants benefited from relatively favorable lending decisions. We trace these disparities to miscalibration in the platform's underwriting model, which underestimates credit risk for Black borrowers and overestimates risk for women. We show that one could correct this miscalibration -- and the corresponding lending disparities -- by explicitly including race and gender in underwriting models, illustrating a tension between competing notions of fairness.
Similar Papers
Price discrimination, algorithmic decision-making, and European non-discrimination law
Computers and Society
Protects people from unfair online prices.
Bridging Research Gaps Between Academic Research and Legal Investigations of Algorithmic Discrimination
Computers and Society
Helps lawyers fight unfair computer decisions.
Bridging Research Gaps Between Academic Research and Legal Investigations of Algorithmic Discrimination
Computers and Society
Helps lawyers fight unfair computer decisions.