ASR-FAIRBENCH: Measuring and Benchmarking Equity Across Speech Recognition Systems
By: Anand Rai , Satyam Rahangdale , Utkarsh Anand and more
Potential Business Impact:
Makes voice assistants work equally for everyone.
Automatic Speech Recognition (ASR) systems have become ubiquitous in everyday applications, yet significant disparities in performance across diverse demographic groups persist. In this work, we introduce the ASR-FAIRBENCH leaderboard which is designed to assess both the accuracy and equity of ASR models in real-time. Leveraging the Meta's Fair-Speech dataset, which captures diverse demographic characteristics, we employ a mixed-effects Poisson regression model to derive an overall fairness score. This score is integrated with traditional metrics like Word Error Rate (WER) to compute the Fairness Adjusted ASR Score (FAAS), providing a comprehensive evaluation framework. Our approach reveals significant performance disparities in SOTA ASR models across demographic groups and offers a benchmark to drive the development of more inclusive ASR technologies.
Similar Papers
How to Evaluate Automatic Speech Recognition: Comparing Different Performance and Bias Measures
Computation and Language
Makes voice assistants work fairly for everyone.
Fairness of Automatic Speech Recognition: Looking Through a Philosophical Lens
Computation and Language
Makes voice assistants understand all accents fairly.
Exploring Gender Disparities in Automatic Speech Recognition Technology
Computation and Language
Makes voice assistants understand everyone equally.