SingLEM: Single-Channel Large EEG Model
By: Jamiyan Sukhbaatar , Satoshi Imamura , Ibuki Inoue and more
Potential Business Impact:
Reads brain signals from one wire.
Current deep learning models for electroencephalography (EEG) are often task-specific and depend on large labeled datasets, limiting their adaptability. Although emerging foundation models aim for broader applicability, their rigid dependence on fixed, high-density multi-channel montages restricts their use across heterogeneous datasets and in missing-channel or practical low-channel settings. To address these limitations, we introduce SingLEM, a self-supervised foundation model that learns robust, general-purpose representations from single-channel EEG, making it inherently hardware agnostic. The model employs a hybrid encoder architecture that combines convolutional layers to extract local features with a hierarchical transformer to model both short- and long-range temporal dependencies. SingLEM is pretrained on 71 public datasets comprising over 9,200 subjects and 357,000 single-channel hours of EEG. When evaluated as a fixed feature extractor across six motor imagery and cognitive tasks, aggregated single-channel representations consistently outperformed leading multi-channel foundation models and handcrafted baselines. These results demonstrate that a single-channel approach can achieve state-of-the-art generalization while enabling fine-grained neurophysiological analysis and enhancing interpretability. The source code and pretrained models are available at https://github.com/ttlabtuat/SingLEM.
Similar Papers
Large Cognition Model: Towards Pretrained EEG Foundation Model
Signal Processing
Reads brain waves to understand thoughts and feelings.
ALFEE: Adaptive Large Foundation Model for EEG Representation
Signal Processing
Helps computers understand brain signals better.
An Efficient Self-Supervised Framework for Long-Sequence EEG Modeling
Machine Learning (CS)
Reads brain signals better and faster.