Convergent transformations of visual representation in brains and models
By: Pablo Marcos-Manchón, Lluís Fuentemilla
Potential Business Impact:
Brain and AI see the world the same way.
A fundamental question in cognitive neuroscience is what shapes visual perception: the external world's structure or the brain's internal architecture. Although some perceptual variability can be traced to individual differences, brain responses to naturalistic stimuli evoke similar activity patterns across individuals, suggesting a convergent representational principle. Here, we test if this stimulus-driven convergence follows a common trajectory across people and deep neural networks (DNNs) during its transformation from sensory to high-level internal representations. We introduce a unified framework that traces representational flow by combining inter-subject similarity with alignment to model hierarchies. Applying this framework to three independent fMRI datasets of visual scene perception, we reveal a cortex-wide network, conserved across individuals, organized into two pathways: a medial-ventral stream for scene structure and a lateral-dorsal stream tuned for social and biological content. This functional organization is captured by the hierarchies of vision DNNs but not language models, reinforcing the specificity of the visual-to-semantic transformation. These findings show a convergent computational solution for visual encoding in both human and artificial vision, driven by the structure of the external world.
Similar Papers
Representations in vision and language converge in a shared, multidimensional space of perceived similarities
Neurons and Cognition
Makes brains and computers understand pictures and words.
Disentangling the Factors of Convergence between Brains and Computer Vision Models
Artificial Intelligence
AI learns to see like humans by changing its training.
Cognitive Neural Architecture Search Reveals Hierarchical Entailment
Neural and Evolutionary Computing
Makes computers see like brains do.