Counterfactual reasoning: an analysis of in-context emergence
By: Moritz Miller, Bernhard Schölkopf, Siyuan Guo
Potential Business Impact:
Helps computers guess what happens if things change.
Large-scale neural language models (LMs) exhibit remarkable performance in in-context learning: the ability to learn and reason the input context on the fly without parameter update. This work studies in-context counterfactual reasoning in language models, that is, to predict the consequences of changes under hypothetical scenarios. We focus on studying a well-defined synthetic setup: a linear regression task that requires noise abduction, where accurate prediction is based on inferring and copying the contextual noise from factual observations. We show that language models are capable of counterfactual reasoning in this controlled setup and provide insights that counterfactual reasoning for a broad class of functions can be reduced to a transformation on in-context observations; we find self-attention, model depth, and data diversity in pre-training drive performance in Transformers. More interestingly, our findings extend beyond regression tasks and show that Transformers can perform noise abduction on sequential data, providing preliminary evidence on the potential for counterfactual story generation. Our code is available under https://github.com/moXmiller/counterfactual-reasoning.git .
Similar Papers
Abstract Counterfactuals for Language Model Agents
Machine Learning (CS)
Helps AI understand "what if" questions better.
On the Eligibility of LLMs for Counterfactual Reasoning: A Decompositional Study
Artificial Intelligence
Helps computers think about "what if" better.
LLMs Struggle to Perform Counterfactual Reasoning with Parametric Knowledge
Artificial Intelligence
Computers can't easily mix old and new facts.