Large Language Models based ASR Error Correction for Child Conversations
By: Anfeng Xu , Tiantian Feng , So Hyun Kim and more
Potential Business Impact:
Makes computers understand kids' talking better.
Automatic Speech Recognition (ASR) has recently shown remarkable progress, but accurately transcribing children's speech remains a significant challenge. Recent developments in Large Language Models (LLMs) have shown promise in improving ASR transcriptions. However, their applications in child speech including conversational scenarios are underexplored. In this study, we explore the use of LLMs in correcting ASR errors for conversational child speech. We demonstrate the promises and challenges of LLMs through experiments on two children's conversational speech datasets with both zero-shot and fine-tuned ASR outputs. We find that while LLMs are helpful in correcting zero-shot ASR outputs and fine-tuned CTC-based ASR outputs, it remains challenging for LLMs to improve ASR performance when incorporating contextual information or when using fine-tuned autoregressive ASR (e.g., Whisper) outputs.
Similar Papers
Customizing Speech Recognition Model with Large Language Model Feedback
Computation and Language
Helps computers understand rare words in speech.
Speech Translation Refinement using Large Language Models
Computation and Language
Makes talking computers translate languages better.
CMT-LLM: Contextual Multi-Talker ASR Utilizing Large Language Models
Audio and Speech Processing
Helps computers understand many people talking at once.