Towards Generalisable Foundation Models for 3D Brain MRI
By: Moona Mazher, Geoff J. M. Parker, Daniel C. Alexander
Potential Business Impact:
Helps doctors find brain problems from scans.
Foundation models in artificial intelligence (AI) are transforming medical imaging by enabling general-purpose feature learning from large-scale, unlabeled datasets. In this work, we introduce BrainFound, a self-supervised foundation model for brain MRI, built by extending DINO-v2, a vision transformer originally designed for 2D natural images. BrainFound adapts DINO-v2 to model full 3D brain anatomy by incorporating volumetric information from sequential MRI slices, moving beyond conventional single-slice paradigms. It supports both single- and multimodal inputs, enabling a broad range of downstream tasks, including disease detection and image segmentation, while generalising across varied imaging protocols and clinical scenarios. We show that BrainFound consistently outperforms existing self-supervised pretraining strategies and supervised baselines, particularly in label-scarce and multi-contrast settings. By integrating information from diverse 3D MRI modalities (e.g., T1, T2, FLAIR), it enhances diagnostic accuracy and reduces dependency on extensive expert annotations. This flexibility makes BrainFound a scalable and practical solution for 3D neuroimaging pipelines, with significant potential for clinical deployment and research innovation.
Similar Papers
MedDINOv3: How to adapt vision foundation models for medical image segmentation?
CV and Pattern Recognition
Helps doctors see organs and sickness in scans.
Building a General SimCLR Self-Supervised Foundation Model Across Neurological Diseases to Advance 3D Brain MRI Diagnoses
CV and Pattern Recognition
Helps doctors find brain problems from scans.
A Foundation Model for Brain MRI with Dynamic Modality Integration
CV and Pattern Recognition
Helps doctors see brain problems with fewer scans.