NAP: Attention-Based Late Fusion for Automatic Sleep Staging
By: Alvise Dei Rossi , Julia van der Meer , Markus H. Schmidt and more
Potential Business Impact:
Helps doctors understand sleep better from brain waves.
Polysomnography signals are highly heterogeneous, varying in modality composition (e.g., EEG, EOG, ECG), channel availability (e.g., frontal, occipital EEG), and acquisition protocols across datasets and clinical sites. Most existing models that process polysomnography data rely on a fixed subset of modalities or channels and therefore neglect to fully exploit its inherently multimodal nature. We address this limitation by introducing NAP (Neural Aggregator of Predictions), an attention-based model which learns to combine multiple prediction streams using a tri-axial attention mechanism that captures temporal, spatial, and predictor-level dependencies. NAP is trained to adapt to different input dimensions. By aggregating outputs from frozen, pretrained single-channel models, NAP consistently outperforms individual predictors and simple ensembles, achieving state-of-the-art zero-shot generalization across multiple datasets. While demonstrated in the context of automated sleep staging from polysomnography, the proposed approach could be extended to other multimodal physiological applications.
Similar Papers
NeuroLingua: A Language-Inspired Hierarchical Framework for Multimodal Sleep Stage Classification Using EEG and EOG
Machine Learning (CS)
Helps machines understand sleep stages better.
On Improving PPG-Based Sleep Staging: A Pilot Study
Signal Processing
Makes smartwatches better at tracking sleep stages.
Stanford Sleep Bench: Evaluating Polysomnography Pre-training Methods for Sleep Foundation Models
Machine Learning (CS)
Helps doctors understand sleep problems better.