Modular Arithmetic: Language Models Solve Math Digit by Digit
By: Tanja Baeumel , Daniil Gurgurov , Yusser al Ghussin and more
Potential Business Impact:
Helps computers do math like humans.
While recent work has begun to uncover the internal strategies that Large Language Models (LLMs) employ for simple arithmetic tasks, a unified understanding of their underlying mechanisms is still lacking. We extend recent findings showing that LLMs represent numbers in a digit-wise manner and present evidence for the existence of digit-position-specific circuits that LLMs use to perform simple arithmetic tasks, i.e. modular subgroups of MLP neurons that operate independently on different digit positions (units, tens, hundreds). Notably, such circuits exist independently of model size and of tokenization strategy, i.e. both for models that encode longer numbers digit-by-digit and as one token. Using Feature Importance and Causal Interventions, we identify and validate the digit-position-specific circuits, revealing a compositional and interpretable structure underlying the solving of arithmetic problems in LLMs. Our interventions selectively alter the model's prediction at targeted digit positions, demonstrating the causal role of digit-position circuits in solving arithmetic tasks.
Similar Papers
Investigating the interaction of linguistic and mathematical reasoning in language models using multilingual number puzzles
Computation and Language
Computers learn math from different number words.
Unravelling the Mechanisms of Manipulating Numbers in Language Models
Computation and Language
Finds how computers make math mistakes.
Addition in Four Movements: Mapping Layer-wise Information Trajectories in LLMs
Artificial Intelligence
Shows how computers learn to do math step-by-step.