Are Language Models Efficient Reasoners? A Perspective from Logic Programming
By: Andreas Opedal , Yanick Zengaffinen , Haruki Shirakami and more
Potential Business Impact:
Teaches computers to solve problems faster.
Modern language models (LMs) exhibit strong deductive reasoning capabilities, yet standard evaluations emphasize correctness while overlooking a key aspect of human-like reasoning: efficiency. In real-world reasoning scenarios, much of the available information is irrelevant, and effective deductive inference requires identifying and ignoring such distractions. We propose a framework for assessing LM reasoning efficiency through the lens of logic programming, introducing a simple method to align proofs written in natural language -- as generated by an LM -- with shortest proofs found by executing the logic program. Efficiency is quantified by measuring how well a model avoids unnecessary inference. Empirically, we construct a dataset of math word problems injected with various number of irrelevant axioms that vary in semantic overlap with the goal theorem. We find that current LMs show marked accuracy declines under such conditions -- even with minimal, domain-consistent distractions -- and the proofs they generate frequently exhibit detours through irrelevant inferences.
Similar Papers
On the Notion that Language Models Reason
Computation and Language
Computers learn by copying patterns, not thinking.
Investigating Language Model Capabilities to Represent and Process Formal Knowledge: A Preliminary Study to Assist Ontology Engineering
Artificial Intelligence
Helps small computers reason better with logic.
Can Large Language Models Learn Formal Logic? A Data-Driven Training and Evaluation Framework
Machine Learning (CS)
Teaches computers to prove math problems correctly.