MEGState: Phoneme Decoding from Magnetoencephalography Signals
By: Shuntaro Suzuki , Chia-Chun Dan Hsu , Yu Tsao and more
Decoding linguistically meaningful representations from non-invasive neural recordings remains a central challenge in neural speech decoding. Among available neuroimaging modalities, magnetoencephalography (MEG) provides a safe and repeatable means of mapping speech-related cortical dynamics, yet its low signal-to-noise ratio and high temporal dimensionality continue to hinder robust decoding. In this work, we introduce MEGState, a novel architecture for phoneme decoding from MEG signals that captures fine-grained cortical responses evoked by auditory stimuli. Extensive experiments on the LibriBrain dataset demonstrate that MEGState consistently surpasses baseline model across multiple evaluation metrics. These findings highlight the potential of MEG-based phoneme decoding as a scalable pathway toward non-invasive brain-computer interfaces for speech.
Similar Papers
Magnetoencephalography (MEG) Based Non-Invasive Chinese Speech Decoding
Audio and Speech Processing
Lets people talk using brain signals.
A Convolutional Framework for Mapping Imagined Auditory MEG into Listened Brain Responses
Signal Processing
Reads minds to control music and speech.
Estimating Brain Activity with High Spatial and Temporal Resolution using a Naturalistic MEG-fMRI Encoding Model
Neurons and Cognition
Maps brain activity with amazing detail.