Enhancing Contrastive Demonstration Selection with Semantic Diversity for Robust In-Context Machine Translation
By: Owen Patterson, Chee Ng
Potential Business Impact:
Teaches computers to translate languages better.
In-Context Learning (ICL) empowers large language models to perform tasks by conditioning on a few input-output examples. However, the performance of ICL is highly sensitive to the selection of these demonstrations. While existing methods focus on similarity or contrastive selection, they often overlook the importance of diversity among the chosen examples. In this paper, we propose DiverseConE (Diversity-Enhanced Contrastive Example Selection), a novel approach for demonstration selection in in-context learning for machine translation. Our method builds upon contrastive selection by incorporating a diversity enhancement step based on embedding space dissimilarity. We conduct extensive experiments on the Llama2-7b model across four language pairs (English-Chinese, Chinese-English, Russian-German, German-Russian) in 1-shot and 3-shot settings, using COMET20 and COMET22 for evaluation. Our results demonstrate that DiverseConE consistently outperforms strong baseline methods, including random selection, BM25, TopK, and a state-of-the-art contrastive selection method. Further analysis, including diversity metrics and human evaluation, validates the effectiveness of our approach and highlights the benefits of considering demonstration diversity for improved translation quality.
Similar Papers
Exploring the Role of Diversity in Example Selection for In-Context Learning
Information Retrieval
Makes AI smarter by picking better examples.
Data-Efficient Biomedical In-Context Learning: A Diversity-Enhanced Submodular Perspective
Computation and Language
Helps AI learn new medical jobs faster.
Selecting Demonstrations for Many-Shot In-Context Learning via Gradient Matching
Computation and Language
Teaches AI better by picking good examples.