Adapting a Segmentation Foundation Model for Medical Image Classification
By: Pengfei Gu , Haoteng Tang , Islam A. Ebeid and more
Potential Business Impact:
Helps doctors find sickness in body pictures.
Recent advancements in foundation models, such as the Segment Anything Model (SAM), have shown strong performance in various vision tasks, particularly image segmentation, due to their impressive zero-shot segmentation capabilities. However, effectively adapting such models for medical image classification is still a less explored topic. In this paper, we introduce a new framework to adapt SAM for medical image classification. First, we utilize the SAM image encoder as a feature extractor to capture segmentation-based features that convey important spatial and contextual details of the image, while freezing its weights to avoid unnecessary overhead during training. Next, we propose a novel Spatially Localized Channel Attention (SLCA) mechanism to compute spatially localized attention weights for the feature maps. The features extracted from SAM's image encoder are processed through SLCA to compute attention weights, which are then integrated into deep learning classification models to enhance their focus on spatially relevant or meaningful regions of the image, thus improving classification performance. Experimental results on three public medical image classification datasets demonstrate the effectiveness and data-efficiency of our approach.
Similar Papers
pFedSAM: Personalized Federated Learning of Segment Anything Model for Medical Image Segmentation
CV and Pattern Recognition
Helps doctors find diseases in scans privately.
Enhancing SAM with Efficient Prompting and Preference Optimization for Semi-supervised Medical Image Segmentation
CV and Pattern Recognition
Finds body parts in medical pictures faster.
SAM2 for Image and Video Segmentation: A Comprehensive Survey
CV and Pattern Recognition
Helps computers perfectly cut out any object from pictures.