Poster: Recognizing Hidden-in-the-Ear Private Key for Reliable Silent Speech Interface Using Multi-Task Learning
By: Xuefu Dong , Liqiang Xu , Lixing He and more
Silent speech interface (SSI) enables hands-free input without audible vocalization, but most SSI systems do not verify speaker identity. We present HEar-ID, which uses consumer active noise-canceling earbuds to capture low-frequency "whisper" audio and high-frequency ultrasonic reflections. Features from both streams pass through a shared encoder, producing embeddings that feed a contrastive branch for user authentication and an SSI head for silent spelling recognition. This design supports decoding of 50 words while reliably rejecting impostors, all on commodity earbuds with a single model. Experiments demonstrate that HEar-ID achieves strong spelling accuracy and robust authentication.
Similar Papers
Wireless Silent Speech Interface Using Multi-Channel Textile EMG Sensors Integrated into Headphones
Human-Computer Interaction
Lets you talk to computers without making noise.
From Silent Signals to Natural Language: A Dual-Stage Transformer-LLM Approach
Computation and Language
Lets computers understand silent talking better.
AISHELL6-whisper: A Chinese Mandarin Audio-visual Whisper Speech Dataset with Speech Recognition Baselines
Audio and Speech Processing
Helps computers understand quiet talking.