RADAR: Enhancing Radiology Report Generation with Supplementary Knowledge Injection
By: Wenjun Hou , Yi Cheng , Kaishuai Xu and more
Potential Business Impact:
Helps doctors write better X-ray reports.
Large language models (LLMs) have demonstrated remarkable capabilities in various domains, including radiology report generation. Previous approaches have attempted to utilize multimodal LLMs for this task, enhancing their performance through the integration of domain-specific knowledge retrieval. However, these approaches often overlook the knowledge already embedded within the LLMs, leading to redundant information integration. To address this limitation, we propose Radar, a framework for enhancing radiology report generation with supplementary knowledge injection. Radar improves report generation by systematically leveraging both the internal knowledge of an LLM and externally retrieved information. Specifically, it first extracts the model's acquired knowledge that aligns with expert image-based classification outputs. It then retrieves relevant supplementary knowledge to further enrich this information. Finally, by aggregating both sources, Radar generates more accurate and informative radiology reports. Extensive experiments on MIMIC-CXR, CheXpert-Plus, and IU X-ray demonstrate that our model outperforms state-of-the-art LLMs in both language quality and clinical accuracy.
Similar Papers
Leveraging LLMs for Multimodal Retrieval-Augmented Radiology Report Generation via Key Phrase Extraction
CV and Pattern Recognition
Helps doctors write X-ray reports faster.
RAD: Towards Trustworthy Retrieval-Augmented Multi-modal Clinical Diagnosis
Machine Learning (CS)
Helps doctors diagnose illnesses using AI.
A Multimodal Multi-Agent Framework for Radiology Report Generation
Artificial Intelligence
Helps doctors write faster, more accurate patient reports.