Unlocking Latent Discourse Translation in LLMs Through Quality-Aware Decoding
By: Wafaa Mohammed, Vlad Niculae, Chrysoula Zerva
Potential Business Impact:
Makes computer translations understand stories better.
Large language models (LLMs) have emerged as strong contenders in machine translation.Yet, they still struggle to adequately handle discourse phenomena, such as pronoun resolution and lexical cohesion at the document level. In this study, we thoroughly investigate the discourse phenomena performance of LLMs in context-aware translation. We demonstrate that discourse knowledge is encoded within LLMs and propose the use of quality-aware decoding (QAD) to effectively extract this knowledge, showcasing its superiority over other decoding approaches through comprehensive analysis. Furthermore, we illustrate that QAD enhances the semantic richness of translations and aligns them more closely with human preferences.
Similar Papers
Multilingual Contextualization of Large Language Models for Document-Level Machine Translation
Computation and Language
Translates whole books, not just sentences.
Beyond the Sentence: A Survey on Context-Aware Machine Translation with Large Language Models
Computation and Language
Makes computer translations understand more context.
Probing LLMs for Multilingual Discourse Generalization Through a Unified Label Set
Computation and Language
Computers understand how sentences connect across languages.