DPO-Tuned Large Language Models for Segmentation in Simultaneous Speech Translation
By: Zeyu Yang, Satoshi Nakamura
Potential Business Impact:
Makes real-time translation sound more natural.
Simultaneous speech translation requires accurate segmentation to balance translation quality and latency. Recent studies such as SHAS have introduced pretrained segmentation models, achieving stronger performance than heuristic rules. However, segmentation models such as SHAS, though pretrained and more robust than heuristic methods, are still constrained by supervised learning objectives and do not incorporate human preference alignment, which is crucial for natural real-time interpretation. In this work, we propose a segmentation framework based on large language models (LLMs) trained with Direct Preference Optimization (DPO). By leveraging preference alignment, our method enables LLMs to predict natural segmentation points that better meet the demands of real-time translation. We evaluate the system on the ACL 60/60 corpus across three language pairs (English-Japanese, Chinese, German), using SeamlessM4T v2 as the translation backbone. Experimental results show that our DPO-tuned LLM achieves higher segmentation accuracy than SHAS and yields consistent improvements in translation quality (BLEU, COMET) as well as latency (Average Lagging). Furthermore, our system benefits from IWSLT baselines for direct comparison. These findings highlight the potential of preference-tuned LLMs to surpass existing pretrained segmentation models and advance adaptive, human-aligned simultaneous interpretation.
Similar Papers
Improving LLMs for Machine Translation Using Synthetic Preference Data
Computation and Language
Makes computer translations much better and more accurate.
Implicit Cross-Lingual Rewarding for Efficient Multilingual Preference Alignment
Computation and Language
Teaches AI to understand more languages from English.
Beyond Single-Reward: Multi-Pair, Multi-Perspective Preference Optimization for Machine Translation
Computation and Language
Teaches computers to translate languages better.