Multilingual Contextualization of Large Language Models for Document-Level Machine Translation
By: Miguel Moura Ramos , Patrick Fernandes , Sweta Agrawal and more
Potential Business Impact:
Translates whole books, not just sentences.
Large language models (LLMs) have demonstrated strong performance in sentence-level machine translation, but scaling to document-level translation remains challenging, particularly in modeling long-range dependencies and discourse phenomena across sentences and paragraphs. In this work, we propose a method to improve LLM-based long-document translation through targeted fine-tuning on high-quality document-level data, which we curate and introduce as DocBlocks. Our approach supports multiple translation paradigms, including direct document-to-document and chunk-level translation, by integrating instructions both with and without surrounding context. This enables models to better capture cross-sentence dependencies while maintaining strong sentence-level translation performance. Experimental results show that incorporating multiple translation paradigms improves document-level translation quality and inference speed compared to prompting and agent-based methods.
Similar Papers
Beyond the Sentence: A Survey on Context-Aware Machine Translation with Large Language Models
Computation and Language
Makes computer translations understand more context.
Two Intermediate Translations Are Better Than One: Fine-tuning LLMs for Document-level Translation Refinement
Computation and Language
Makes translated documents sound more natural.
Improving LLM-based Document-level Machine Translation with Multi-Knowledge Fusion
Computation and Language
Improves computer translation by using summaries and key words.