Propositional Logic for Probing Generalization in Neural Networks
By: Anna Langedijk, Jaap Jumelet, Willem Zuidema
Potential Business Impact:
Computers struggle to learn logic rules.
The extent to which neural networks are able to acquire and represent symbolic rules remains a key topic of research and debate. Much current work focuses on the impressive capabilities of large language models, as well as their often ill-understood failures on a wide range of reasoning tasks. In this paper, in contrast, we investigate the generalization behavior of three key neural architectures (Transformers, Graph Convolution Networks and LSTMs) in a controlled task rooted in propositional logic. The task requires models to generate satisfying assignments for logical formulas, making it a structured and interpretable setting for studying compositionality. We introduce a balanced extension of an existing dataset to eliminate superficial patterns and enable testing on unseen operator combinations. Using this dataset, we evaluate the ability of the three architectures to generalize beyond the training distribution. While all models perform well in-distribution, we find that generalization to unseen patterns, particularly those involving negation, remains a significant challenge. Transformers fail to apply negation compositionally, unless structural biases are introduced. Our findings highlight persistent limitations in the ability of standard architectures to learn systematic representations of logical operators, suggesting the need for stronger inductive biases to support robust rule-based reasoning.
Similar Papers
Robustness of Neurosymbolic Reasoners on First-Order Logic Problems
Computation and Language
Makes computers reason better with tricky logic.
Analyzing Generalization in Pre-Trained Symbolic Regression
Machine Learning (CS)
Finds math formulas that work for new problems.
Hybrid Models for Natural Language Reasoning: The Case of Syllogistic Logic
Computation and Language
Helps computers reason logically by combining two skills.