Extending Automatic Machine Translation Evaluation to Book-Length Documents
By: Kuang-Da Wang , Shuoyang Ding , Chao-Han Huck Yang and more
Potential Business Impact:
Tests if computers translate whole books well.
Despite Large Language Models (LLMs) demonstrating superior translation performance and long-context capabilities, evaluation methodologies remain constrained to sentence-level assessment due to dataset limitations, token number restrictions in metrics, and rigid sentence boundary requirements. We introduce SEGALE, an evaluation scheme that extends existing automatic metrics to long-document translation by treating documents as continuous text and applying sentence segmentation and alignment methods. Our approach enables previously unattainable document-level evaluation, handling translations of arbitrary length generated with document-level prompts while accounting for under-/over-translations and varied sentence boundaries. Experiments show our scheme significantly outperforms existing long-form document evaluation schemes, while being comparable to evaluations performed with groundtruth sentence alignments. Additionally, we apply our scheme to book-length texts and newly demonstrate that many open-weight LLMs fail to effectively translate documents at their reported maximum context lengths.
Similar Papers
Same evaluation, more tokens: On the effect of input length for machine translation evaluation using Large Language Models
Computation and Language
Helps computers judge long translations better.
Align-then-Slide: A complete evaluation framework for Ultra-Long Document-Level Machine Translation
Computation and Language
Checks if long translations are good.
Automatic Evaluation Metrics for Document-level Translation: Overview, Challenges and Trends
Computation and Language
Checks if computer translations are good.