The Few-shot Dilemma: Over-prompting Large Language Models
By: Yongjian Tang , Doruk Tuncel , Christian Koerner and more
Potential Business Impact:
Helps AI understand better with fewer examples.
Over-prompting, a phenomenon where excessive examples in prompts lead to diminished performance in Large Language Models (LLMs), challenges the conventional wisdom about in-context few-shot learning. To investigate this few-shot dilemma, we outline a prompting framework that leverages three standard few-shot selection methods - random sampling, semantic embedding, and TF-IDF vectors - and evaluate these methods across multiple LLMs, including GPT-4o, GPT-3.5-turbo, DeepSeek-V3, Gemma-3, LLaMA-3.1, LLaMA-3.2, and Mistral. Our experimental results reveal that incorporating excessive domain-specific examples into prompts can paradoxically degrade performance in certain LLMs, which contradicts the prior empirical conclusion that more relevant few-shot examples universally benefit LLMs. Given the trend of LLM-assisted software engineering and requirement analysis, we experiment with two real-world software requirement classification datasets. By gradually increasing the number of TF-IDF-selected and stratified few-shot examples, we identify their optimal quantity for each LLM. This combined approach achieves superior performance with fewer examples, avoiding the over-prompting problem, thus surpassing the state-of-the-art by 1% in classifying functional and non-functional requirements.
Similar Papers
When Many-Shot Prompting Fails: An Empirical Study of LLM Code Translation
Software Engineering
Few examples help computers translate code best.
Retrieval-Augmented Few-Shot Prompting Versus Fine-Tuning for Code Vulnerability Detection
Software Engineering
Finds computer code flaws better without retraining.
Instructional Prompt Optimization for Few-Shot LLM-Based Recommendations on Cold-Start Users
Artificial Intelligence
Helps new users get good suggestions faster.