Medical AI Consensus: A Multi-Agent Framework for Radiology Report Generation and Evaluation
By: Ahmed T. Elboardy, Ghada Khoriba, Essam A. Rashed
Potential Business Impact:
Helps doctors write patient reports faster.
Automating radiology report generation poses a dual challenge: building clinically reliable systems and designing rigorous evaluation protocols. We introduce a multi-agent reinforcement learning framework that serves as both a benchmark and evaluation environment for multimodal clinical reasoning in the radiology ecosystem. The proposed framework integrates large language models (LLMs) and large vision models (LVMs) within a modular architecture composed of ten specialized agents responsible for image analysis, feature extraction, report generation, review, and evaluation. This design enables fine-grained assessment at both the agent level (e.g., detection and segmentation accuracy) and the consensus level (e.g., report quality and clinical relevance). We demonstrate an implementation using chatGPT-4o on public radiology datasets, where LLMs act as evaluators alongside medical radiologist feedback. By aligning evaluation protocols with the LLM development lifecycle, including pretraining, finetuning, alignment, and deployment, the proposed benchmark establishes a path toward trustworthy deviance-based radiology report generation.
Similar Papers
A Multimodal Multi-Agent Framework for Radiology Report Generation
Artificial Intelligence
Helps doctors write faster, more accurate patient reports.
Aligning Findings with Diagnosis: A Self-Consistent Reinforcement Learning Framework for Trustworthy Radiology Reporting
Machine Learning (CS)
Helps doctors write accurate patient reports faster.
MRGAgents: A Multi-Agent Framework for Improved Medical Report Generation with Med-LVLMs
Multiagent Systems
Helps doctors find hidden problems in X-rays.