Score: 0

Comparing Large Language Models and Traditional Machine Translation Tools for Translating Medical Consultation Summaries: A Pilot Study

Published: April 23, 2025 | arXiv ID: 2504.16601v1

By: Andy Li , Wei Zhou , Rashina Hoda and more

Potential Business Impact:

Helps doctors translate patient notes better.

Business Areas:
Translation Service Professional Services

This study evaluates how well large language models (LLMs) and traditional machine translation (MT) tools translate medical consultation summaries from English into Arabic, Chinese, and Vietnamese. It assesses both patient, friendly and clinician, focused texts using standard automated metrics. Results showed that traditional MT tools generally performed better, especially for complex texts, while LLMs showed promise, particularly in Vietnamese and Chinese, when translating simpler summaries. Arabic translations improved with complexity due to the language's morphology. Overall, while LLMs offer contextual flexibility, they remain inconsistent, and current evaluation metrics fail to capture clinical relevance. The study highlights the need for domain-specific training, improved evaluation methods, and human oversight in medical translation.

Country of Origin
🇦🇺 Australia

Page Count
8 pages

Category
Computer Science:
Computation and Language