Hybrid Deep Learning and Handcrafted Feature Fusion for Mammographic Breast Cancer Classification
By: Maximilian Tschuchnig, Michael Gadermayr, Khalifa Djemal
Potential Business Impact:
Helps doctors find breast cancer faster.
Automated breast cancer classification from mammography remains a significant challenge due to subtle distinctions between benign and malignant tissue. In this work, we present a hybrid framework combining deep convolutional features from a ResNet-50 backbone with handcrafted descriptors and transformer-based embeddings. Using the CBIS-DDSM dataset, we benchmark our ResNet-50 baseline (AUC: 78.1%) and demonstrate that fusing handcrafted features with deep ResNet-50 and DINOv2 features improves AUC to 79.6% (setup d1), with a peak recall of 80.5% (setup d1) and highest F1 score of 67.4% (setup d1). Our experiments show that handcrafted features not only complement deep representations but also enhance performance beyond transformer-based embeddings. This hybrid fusion approach achieves results comparable to state-of-the-art methods while maintaining architectural simplicity and computational efficiency, making it a practical and effective solution for clinical decision support.
Similar Papers
Attention-Enhanced Deep Learning Ensemble for Breast Density Classification in Mammography
Image and Video Processing
Finds dense breasts to spot cancer early.
Enhancing breast cancer detection on screening mammogram using self-supervised learning and a hybrid deep model of Swin Transformer and Convolutional Neural Network
CV and Pattern Recognition
Helps doctors find breast cancer on X-rays.
A Hybrid CNN-VSSM model for Multi-View, Multi-Task Mammography Analysis: Robust Diagnosis with Attention-Based Fusion
Image and Video Processing
Finds breast cancer earlier and more accurately.