Learning from Generalization Patterns: An Evaluation-Driven Approach to Enhanced Data Augmentation for Fine-Tuning Small Language Models
By: Huan Song , Deeksha Razdan , Yiyue Qian and more
Potential Business Impact:
Makes small AI smarter with less work.
Small Language Models (SLMs) offer compelling advantages in deployment cost and latency, but their accuracy often lags behind larger models, particularly for complex domain-specific tasks. While supervised fine-tuning can help bridge this performance gap, it requires substantial manual effort in data preparation and iterative optimization. We present PaDA-Agent (Pattern-guided Data Augmentation Agent), an evaluation-driven approach that streamlines the data augmentation process for SLMs through coordinated operations. Unlike state-of-the-art approaches that focus on model training errors only and generating error-correcting samples, PaDA-Agent discovers failure patterns from the validation data via evaluations and drafts targeted data augmentation strategies aiming to directly reduce the generalization gap. Our experimental results demonstrate significant improvements over state-of-the-art LLM-based data augmentation approaches for Llama 3.2 1B Instruct model fine-tuning.
Similar Papers
Self-Improving LLM Agents at Test-Time
Machine Learning (CS)
Computers learn better by fixing their own mistakes.
SALAD: Improving Robustness and Generalization through Contrastive Learning with Structure-Aware and LLM-Driven Augmented Data
Computation and Language
Teaches computers to understand language better.
AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs
Machine Learning (CS)
Lets AI learn new things without retraining.