Scalable Offline ASR for Command-Style Dictation in Courtrooms
By: Kumarmanas Nethil , Vaibhav Mishra , Kriti Anandan and more
Potential Business Impact:
Lets many people talk to computers at once.
We propose an open-source framework for Command-style dictation that addresses the gap between resource-intensive Online systems and high-latency Batch processing. Our approach uses Voice Activity Detection (VAD) to segment audio and transcribes these segments in parallel using Whisper models, enabling efficient multiplexing across audios. Unlike proprietary systems like SuperWhisper, this framework is also compatible with most ASR architectures, including widely used CTC-based models. Our multiplexing technique maximizes compute utilization in real-world settings, as demonstrated by its deployment in around 15% of India's courtrooms. Evaluations on live data show consistent latency reduction as user concurrency increases, compared to sequential batch processing. The live demonstration will showcase our open-sourced implementation and allow attendees to interact with it in real-time.
Similar Papers
WhisperKit: On-device Real-time ASR with Billion-Scale Transformers
Sound
Lets phones understand your voice super fast.
LibriVAD: A Scalable Open Dataset with Deep Learning Benchmarks for Voice Activity Detection
Sound
Helps computers hear talking in loud places.
AS-ASR: A Lightweight Framework for Aphasia-Specific Automatic Speech Recognition
Audio and Speech Processing
Helps people with speech problems talk to computers.