A MIND for Reasoning: Meta-learning for In-context Deduction
By: Leonardo Bertolazzi , Manuel Vargas Guzmán , Raffaella Bernardi and more
Potential Business Impact:
Helps small AI learn to reason better.
Large language models (LLMs) are increasingly evaluated on formal tasks, where strong reasoning abilities define the state of the art. However, their ability to generalize to out-of-distribution problems remains limited. In this paper, we investigate how LLMs can achieve a systematic understanding of deductive rules. Our focus is on the task of identifying the appropriate subset of premises within a knowledge base needed to derive a given hypothesis. To tackle this challenge, we propose Meta-learning for In-context Deduction (MIND), a novel few-shot meta-learning fine-tuning approach. The goal of MIND is to enable models to generalize more effectively to unseen knowledge bases and to systematically apply inference rules. Our results show that MIND significantly improves generalization in small LMs ranging from 1.5B to 7B parameters. The benefits are especially pronounced in smaller models and low-data settings. Remarkably, small models fine-tuned with MIND outperform state-of-the-art LLMs, such as GPT-4o and o3-mini, on this task.
Similar Papers
MIND: From Passive Mimicry to Active Reasoning through Capability-Aware Multi-Perspective CoT Distillation
Computation and Language
Teaches small computers big thinking skills.
Meta-Reasoner: Dynamic Guidance for Optimized Inference-time Reasoning in Large Language Models
Artificial Intelligence
Helps computers solve problems smarter and faster.
From Reasoning to Learning: A Survey on Hypothesis Discovery and Rule Learning with Large Language Models
Artificial Intelligence
Computers can now invent new ideas and theories.