Faithful Summarization of Consumer Health Queries: A Cross-Lingual Framework with LLMs
By: Ajwad Abrar , Nafisa Tabassum Oeshy , Prianka Maheru and more
Potential Business Impact:
Makes doctor notes easier to understand safely.
Summarizing consumer health questions (CHQs) can ease communication in healthcare, but unfaithful summaries that misrepresent medical details pose serious risks. We propose a framework that combines TextRank-based sentence extraction and medical named entity recognition with large language models (LLMs) to enhance faithfulness in medical text summarization. In our experiments, we fine-tuned the LLaMA-2-7B model on the MeQSum (English) and BanglaCHQ-Summ (Bangla) datasets, achieving consistent improvements across quality (ROUGE, BERTScore, readability) and faithfulness (SummaC, AlignScore) metrics, and outperforming zero-shot baselines and prior systems. Human evaluation further shows that over 80\% of generated summaries preserve critical medical information. These results highlight faithfulness as an essential dimension for reliable medical summarization and demonstrate the potential of our approach for safer deployment of LLMs in healthcare contexts.
Similar Papers
Performance Evaluation of Large Language Models in Bangla Consumer Health Query Summarization
Computation and Language
Helps computers understand health questions in Bengali.
Topic-aware Large Language Models for Summarizing the Lived Healthcare Experiences Described in Health Stories
Computers and Society
Helps doctors understand patient stories better.
Disparities in Multilingual LLM-Based Healthcare Q&A
Computation and Language
Makes AI give fair health answers in any language.