Advanced Chest X-Ray Analysis via Transformer-Based Image Descriptors and Cross-Model Attention Mechanism
By: Lakshita Agarwal, Bindu Verma
Potential Business Impact:
Reads X-rays and explains what's wrong.
The examination of chest X-ray images is a crucial component in detecting various thoracic illnesses. This study introduces a new image description generation model that integrates a Vision Transformer (ViT) encoder with cross-modal attention and a GPT-4-based transformer decoder. The ViT captures high-quality visual features from chest X-rays, which are fused with text data through cross-modal attention to improve the accuracy, context, and richness of image descriptions. The GPT-4 decoder transforms these fused features into accurate and relevant captions. The model was tested on the National Institutes of Health (NIH) and Indiana University (IU) Chest X-ray datasets. On the IU dataset, it achieved scores of 0.854 (B-1), 0.883 (CIDEr), 0.759 (METEOR), and 0.712 (ROUGE-L). On the NIH dataset, it achieved the best performance on all metrics: BLEU 1--4 (0.825, 0.788, 0.765, 0.752), CIDEr (0.857), METEOR (0.726), and ROUGE-L (0.705). This framework has the potential to enhance chest X-ray evaluation, assisting radiologists in more precise and efficient diagnosis.
Similar Papers
Automated diagnosis of lung diseases using vision transformer: a comparative study on chest x-ray classification
Image and Video Processing
Finds pneumonia on X-rays with 99% accuracy.
Beyond Conventional Transformers: The Medical X-ray Attention (MXA) Block for Improved Multi-Label Diagnosis Using Knowledge Distillation
CV and Pattern Recognition
Helps doctors find more sicknesses on X-rays.
MedDChest: A Content-Aware Multimodal Foundational Vision Model for Thoracic Imaging
CV and Pattern Recognition
Helps doctors find sickness in chest scans.