Contextual Cues in Machine Translation: Investigating the Potential of Multi-Source Input Strategies in LLMs and NMT Systems
By: Lia Shahnazaryan, Patrick Simianer, Joern Wuebker
Potential Business Impact:
Improves computer translations by adding extra clues.
We explore the impact of multi-source input strategies on machine translation (MT) quality, comparing GPT-4o, a large language model (LLM), with a traditional multilingual neural machine translation (NMT) system. Using intermediate language translations as contextual cues, we evaluate their effectiveness in enhancing English and Chinese translations into Portuguese. Results suggest that contextual information significantly improves translation quality for domain-specific datasets and potentially for linguistically distant language pairs, with diminishing returns observed in benchmarks with high linguistic variability. Additionally, we demonstrate that shallow fusion, a multi-source approach we apply within the NMT system, shows improved results when using high-resource languages as context for other translation pairs, highlighting the importance of strategic context language selection.
Similar Papers
Beyond the Sentence: A Survey on Context-Aware Machine Translation with Large Language Models
Computation and Language
Makes computer translations understand more context.
Improving LLM-based Document-level Machine Translation with Multi-Knowledge Fusion
Computation and Language
Improves computer translation by using summaries and key words.
Context-Aware Monolingual Human Evaluation of Machine Translation
Computation and Language
Lets people check translations without the original text.