Exploring the Role of Diversity in Example Selection for In-Context Learning
By: Janak Kapuriya , Manit Kaushik , Debasis Ganguly and more
Potential Business Impact:
Makes AI smarter by picking better examples.
In-Context Learning (ICL) has gained prominence due to its ability to perform tasks without requiring extensive training data and its robustness to noisy labels. A typical ICL workflow involves selecting localized examples relevant to a given input using sparse or dense embedding-based similarity functions. However, relying solely on similarity-based selection may introduce topical biases in the retrieved contexts, potentially leading to suboptimal downstream performance. We posit that reranking the retrieved context to enhance topical diversity can improve downstream task performance. To achieve this, we leverage maximum marginal relevance (MMR) which balances topical similarity with inter-example diversity. Our experimental results demonstrate that diversifying the selected examples leads to consistent improvements in downstream performance across various context sizes and similarity functions. The implementation of our approach is made available at https://github.com/janak11111/Diverse-ICL.
Similar Papers
Enhancing Contrastive Demonstration Selection with Semantic Diversity for Robust In-Context Machine Translation
Computation and Language
Teaches computers to translate languages better.
KITE: Kernelized and Information Theoretic Exemplars for In-Context Learning
Machine Learning (CS)
Picks best examples to help AI answer questions.
Refract ICL: Rethinking Example Selection in the Era of Million-Token Models
Computation and Language
Teaches AI to learn better from many examples.