Score: 0

DPO-Tuned Large Language Models for Segmentation in Simultaneous Speech Translation

Published: October 14, 2025 | arXiv ID: 2510.12195v1

By: Zeyu Yang, Satoshi Nakamura

Potential Business Impact:

Makes real-time translation sound more natural.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Simultaneous speech translation requires accurate segmentation to balance translation quality and latency. Recent studies such as SHAS have introduced pretrained segmentation models, achieving stronger performance than heuristic rules. However, segmentation models such as SHAS, though pretrained and more robust than heuristic methods, are still constrained by supervised learning objectives and do not incorporate human preference alignment, which is crucial for natural real-time interpretation. In this work, we propose a segmentation framework based on large language models (LLMs) trained with Direct Preference Optimization (DPO). By leveraging preference alignment, our method enables LLMs to predict natural segmentation points that better meet the demands of real-time translation. We evaluate the system on the ACL 60/60 corpus across three language pairs (English-Japanese, Chinese, German), using SeamlessM4T v2 as the translation backbone. Experimental results show that our DPO-tuned LLM achieves higher segmentation accuracy than SHAS and yields consistent improvements in translation quality (BLEU, COMET) as well as latency (Average Lagging). Furthermore, our system benefits from IWSLT baselines for direct comparison. These findings highlight the potential of preference-tuned LLMs to surpass existing pretrained segmentation models and advance adaptive, human-aligned simultaneous interpretation.

Country of Origin
🇭🇰 Hong Kong

Page Count
5 pages

Category
Computer Science:
Computation and Language