Score: 0

Real-Time Sign Language Gestures to Speech Transcription using Deep Learning

Published: August 18, 2025 | arXiv ID: 2508.12713v1

By: Brandone Fonya

Potential Business Impact:

Translates sign language into speech instantly.

Communication barriers pose significant challenges for individuals with hearing and speech impairments, often limiting their ability to effectively interact in everyday environments. This project introduces a real-time assistive technology solution that leverages advanced deep learning techniques to translate sign language gestures into textual and audible speech. By employing convolution neural networks (CNN) trained on the Sign Language MNIST dataset, the system accurately classifies hand gestures captured live via webcam. Detected gestures are instantaneously translated into their corresponding meanings and transcribed into spoken language using text-to-speech synthesis, thus facilitating seamless communication. Comprehensive experiments demonstrate high model accuracy and robust real-time performance with some latency, highlighting the system's practical applicability as an accessible, reliable, and user-friendly tool for enhancing the autonomy and integration of sign language users in diverse social settings.

Country of Origin
🇺🇸 United States

Page Count
6 pages

Category
Computer Science:
CV and Pattern Recognition