Rethinking Label Consistency of In-Context Learning: An Implicit Transductive Label Propagation Perspective
By: Haoyang Chen, Richong Zhang, Junfan Chen
Large language models (LLMs) perform in-context learning (ICL) with minimal supervised examples, which benefits various natural language processing (NLP) tasks. One of the critical research focus is the selection of prompt demonstrations. Current approaches typically employ retrieval models to select the top-K most semantically similar examples as demonstrations. However, we argue that existing methods are limited since the label consistency is not guaranteed during demonstration selection. Our cognition derives from the Bayesian view of ICL and our rethinking of ICL from the transductive label propagation perspective. We treat ICL as a transductive learning method and incorporate latent concepts from Bayesian view and deduce that similar demonstrations guide the concepts of query, with consistent labels serving as estimates. Based on this understanding, we establish a label propagation framework to link label consistency with propagation error bounds. To model label consistency, we propose a data synthesis method, leveraging both semantic and label information, and use TopK sampling with Synthetic Data (TopK-SD) to acquire demonstrations with consistent labels. TopK-SD outperforms original TopK sampling on multiple benchmarks. Our work provides a new perspective for understanding the working mechanisms within ICL.
Similar Papers
From Cross-Task Examples to In-Task Prompts: A Graph-Based Pseudo-Labeling Framework for In-context Learning
Artificial Intelligence
Teaches computers new things with fewer examples.
On the Relationship Between the Choice of Representation and In-Context Learning
Computation and Language
Lets computers learn new things better.
Learn to Select: Exploring Label Distribution Divergence for In-Context Demonstration Selection in Text Classification
Computation and Language
Picks best examples to teach computers faster.