Causal-Enhanced AI Agents for Medical Research Screening
By: Duc Ngo, Arya Rahgoza
Potential Business Impact:
AI accurately finds medical facts, no mistakes.
Systematic reviews are essential for evidence-based medicine, but reviewing 1.5 million+ annual publications manually is infeasible. Current AI approaches suffer from hallucinations in systematic review tasks, with studies reporting rates ranging from 28--40% for earlier models to 2--15% for modern implementations which is unacceptable when errors impact patient care. We present a causal graph-enhanced retrieval-augmented generation system integrating explicit causal reasoning with dual-level knowledge graphs. Our approach enforces evidence-first protocols where every causal claim traces to retrieved literature and automatically generates directed acyclic graphs visualizing intervention-outcome pathways. Evaluation on 234 dementia exercise abstracts shows CausalAgent achieves 95% accuracy, 100% retrieval success, and zero hallucinations versus 34% accuracy and 10% hallucinations for baseline AI. Automatic causal graphs enable explicit mechanism modeling, visual synthesis, and enhanced interpretability. While this proof-of-concept evaluation used ten questions focused on dementia exercise research, the architectural approach demonstrates transferable principles for trustworthy medical AI and causal reasoning's potential for high-stakes healthcare.
Similar Papers
Retrieval-augmented systems can be dangerous medical communicators
Computation and Language
AI answers about health can still mislead patients.
CausalGuard: A Smart System for Detecting and Preventing False Information in Large Language Models
Artificial Intelligence
Stops AI from making up wrong answers.
Learning to Seek Evidence: A Verifiable Reasoning Agent with Causal Faithfulness Analysis
Artificial Intelligence
AI explains medical guesses using proof.