Retrieval-Augmented Generation for Electrocardiogram-Language Models
By: Xiaoyu Song , William Han , Tony Chen and more
Potential Business Impact:
Helps doctors understand heart signals better.
Interest in generative Electrocardiogram-Language Models (ELMs) is growing, as they can produce textual responses conditioned on ECG signals and textual queries. Unlike traditional classifiers that output label probabilities, ELMs are more versatile, supporting domain-specific tasks (e.g., waveform analysis, diagnosis, prognosis) as well as general tasks (e.g., open-ended questions, dialogue). Retrieval-Augmented Generation (RAG), widely used in Large Language Models (LLMs) to ground LLM outputs in retrieved knowledge, helps reduce hallucinations and improve natural language generation (NLG). However, despite its promise, no open-source implementation or systematic study of RAG pipeline design for ELMs currently exists. To address this gap, we present the first open-source RAG pipeline for ELMs, along with baselines and ablation studies for NLG. Experiments on three public datasets show that ELMs with RAG consistently improves performance over non-RAG baselines and highlights key ELM design considerations. Our code is available at: https://github.com/willxxy/ECG-Bench.
Similar Papers
ALFRED: Ask a Large-language model For Reliable ECG Diagnosis
Machine Learning (CS)
Helps doctors diagnose heart problems from ECGs.
ECG-LLM -- training and evaluation of domain-specific large language models for electrocardiography
Computation and Language
Helps doctors understand heart signals better.
Evaluating Retrieval-Augmented Generation vs. Long-Context Input for Clinical Reasoning over EHRs
Computation and Language
Helps doctors quickly find patient info in records.