Mistake Notebook Learning: Selective Batch-Wise Context Optimization for In-Context Learning
By: Xuanbo Su , Yingfang Zhang , Hao Luo and more
Potential Business Impact:
Teaches AI to learn from its mistakes.
Large language models (LLMs) adapt to tasks via gradient fine-tuning (heavy computation, catastrophic forgetting) or In-Context Learning (ICL: low robustness, poor mistake learning). To fix this, we introduce Mistake Notebook Learning (MNL), a training-free framework with a persistent knowledge base of abstracted error patterns. Unlike prior instance/single-trajectory memory methods, MNL uses batch-wise error abstraction: it extracts generalizable guidance from multiple failures, stores insights in a dynamic notebook, and retains only baseline-outperforming guidance via hold-out validation (ensuring monotonic improvement). We show MNL nearly matches Supervised Fine-Tuning (93.9% vs 94.3% on GSM8K) and outperforms training-free alternatives on GSM8K, Spider, AIME, and KaggleDBQA. On KaggleDBQA (Qwen3-8B), MNL hits 28% accuracy (47% relative gain), outperforming Memento (15.1%) and Training-Free GRPO (22.1) - proving it's a strong training-free alternative for complex reasoning.
Similar Papers
You Only Fine-tune Once: Many-Shot In-Context Fine-Tuning for Large Language Model
Computation and Language
Teaches computers to do many jobs well at once.
Retrieval Enhanced Feedback via In-context Neural Error-book
Machine Learning (CS)
Teaches AI to learn from mistakes better.
Retrieval Enhanced Feedback via In-context Neural Error-book
Machine Learning (CS)
Teaches AI to learn from mistakes better.