Interpretable EEG-to-Image Generation with Semantic Prompts
By: Arshak Rezvani , Ali Akbari , Kosar Sanjar Arani and more
Potential Business Impact:
Lets computers guess what you see from brain waves.
Decoding visual experience from brain signals offers exciting possibilities for neuroscience and interpretable AI. While EEG is accessible and temporally precise, its limitations in spatial detail hinder image reconstruction. Our model bypasses direct EEG-to-image generation by aligning EEG signals with multilevel semantic captions -- ranging from object-level to abstract themes -- generated by a large language model. A transformer-based EEG encoder maps brain activity to these captions through contrastive learning. During inference, caption embeddings retrieved via projection heads condition a pretrained latent diffusion model for image generation. This text-mediated framework yields state-of-the-art visual decoding on the EEGCVPR dataset, with interpretable alignment to known neurocognitive pathways. Dominant EEG-caption associations reflected the importance of different semantic levels extracted from perceived images. Saliency maps and t-SNE projections reveal semantic topography across the scalp. Our model demonstrates how structured semantic mediation enables cognitively aligned visual decoding from EEG.
Similar Papers
Mind2Matter: Creating 3D Models from EEG Signals
Graphics
Reads minds to build 3D objects.
ViEEG: Hierarchical Visual Neural Representation for EEG Brain Decoding
CV and Pattern Recognition
Lets computers see what you see.
NeuroBridge: Bio-Inspired Self-Supervised EEG-to-Image Decoding via Cognitive Priors and Bidirectional Semantic Alignment
CV and Pattern Recognition
Lets computers see what you see from brain waves.