Trans-Zero: Self-Play Incentivizes Large Language Models for Multilingual Translation Without Parallel Data
By: Wei Zou , Sen Yang , Yu Bao and more
Potential Business Impact:
Translates languages without needing example sentences.
The rise of Large Language Models (LLMs) has reshaped machine translation (MT), but multilingual MT still relies heavily on parallel data for supervised fine-tuning (SFT), facing challenges like data scarcity for low-resource languages and catastrophic forgetting. To address these issues, we propose TRANS-ZERO, a self-play framework that leverages only monolingual data and the intrinsic multilingual knowledge of LLM. TRANS-ZERO combines Genetic Monte-Carlo Tree Search (G-MCTS) with preference optimization, achieving strong translation performance that rivals supervised methods. Experiments demonstrate that this approach not only matches the performance of models trained on large-scale parallel data but also excels in non-English translation directions. Further analysis reveals that G-MCTS itself significantly enhances translation quality by exploring semantically consistent candidates through iterative translations, providing a robust foundation for the framework's succuss.
Similar Papers
MT-R1-Zero: Advancing LLM-based Machine Translation via R1-Zero-like Reinforcement Learning
Computation and Language
Makes computer translations better without needing examples.
Beyond English: Toward Inclusive and Scalable Multilingual Machine Translation with LLMs
Computation and Language
Translates 60 languages better, even Chinese.
Zero-Shot Cross-Lingual Transfer using Prefix-Based Adaptation
Computation and Language
Helps computers learn many languages faster.