Sign Language Recognition using Bidirectional Reservoir Computing
By: Nitin Kumar Singh , Arie Rachmad Syulistyo , Yuichiro Tanaka and more
Potential Business Impact:
Lets phones understand sign language fast.
Sign language recognition (SLR) facilitates communication between deaf and hearing individuals. Deep learning is widely used to develop SLR-based systems; however, it is computationally intensive and requires substantial computational resources, making it unsuitable for resource-constrained devices. To address this, we propose an efficient sign language recognition system using MediaPipe and an echo state network (ESN)-based bidirectional reservoir computing (BRC) architecture. MediaPipe extracts hand joint coordinates, which serve as inputs to the ESN-based BRC architecture. The BRC processes these features in both forward and backward directions, efficiently capturing temporal dependencies. The resulting states of BRC are concatenated to form a robust representation for classification. We evaluated our method on the Word-Level American Sign Language (WLASL) video dataset, achieving a competitive accuracy of 57.71% and a significantly lower training time of only 9 seconds, in contrast to the 55 minutes and $38$ seconds required by the deep learning-based Bi-GRU approach. Consequently, the BRC-based SLR system is well-suited for edge devices.
Similar Papers
SLRNet: A Real-Time LSTM-Based Sign Language Recognition System
CV and Pattern Recognition
Lets computers understand sign language from your webcam.
Isolated Sign Language Recognition with Segmentation and Pose Estimation
CV and Pattern Recognition
Helps computers understand sign language from videos.
Data-Efficient American Sign Language Recognition via Few-Shot Prototypical Networks
CV and Pattern Recognition
Teaches computers to understand rare sign language.