EEG-Driven Image Reconstruction with Saliency-Guided Diffusion Models
By: Igor Abramov, Ilya Makarov
Potential Business Impact:
Shows what you're thinking by drawing pictures.
Existing EEG-driven image reconstruction methods often overlook spatial attention mechanisms, limiting fidelity and semantic coherence. To address this, we propose a dual-conditioning framework that combines EEG embeddings with spatial saliency maps to enhance image generation. Our approach leverages the Adaptive Thinking Mapper (ATM) for EEG feature extraction and fine-tunes Stable Diffusion 2.1 via Low-Rank Adaptation (LoRA) to align neural signals with visual semantics, while a ControlNet branch conditions generation on saliency maps for spatial control. Evaluated on THINGS-EEG, our method achieves a significant improvement in the quality of low- and high-level image features over existing approaches. Simultaneously, strongly aligning with human visual attention. The results demonstrate that attentional priors resolve EEG ambiguities, enabling high-fidelity reconstructions with applications in medical diagnostics and neuroadaptive interfaces, advancing neural decoding through efficient adaptation of pre-trained diffusion models.
Similar Papers
Step-Aware Residual-Guided Diffusion for EEG Spatial Super-Resolution
Multimedia
Makes brain signals clearer for computers.
DynaMind: Reconstructing Dynamic Visual Scenes from EEG by Aligning Temporal Dynamics and Multimodal Semantics to Guided Diffusion
CV and Pattern Recognition
Shows what you're seeing from brain waves.
SAGE: Saliency-Guided Contrastive Embeddings
CV and Pattern Recognition
Teaches computers to see what humans see.