Score: 2

Disaggregation Reveals Hidden Training Dynamics: The Case of Agreement Attraction

Published: October 28, 2025 | arXiv ID: 2510.24934v1

By: James A. Michaelov, Catherine Arnett

BigTech Affiliations: Massachusetts Institute of Technology

Potential Business Impact:

Makes computers learn grammar like kids do.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Language models generally produce grammatical text, but they are more likely to make errors in certain contexts. Drawing on paradigms from psycholinguistics, we carry out a fine-grained analysis of those errors in different syntactic contexts. We demonstrate that by disaggregating over the conditions of carefully constructed datasets and comparing model performance on each over the course of training, it is possible to better understand the intermediate stages of grammatical learning in language models. Specifically, we identify distinct phases of training where language model behavior aligns with specific heuristics such as word frequency and local context rather than generalized grammatical rules. We argue that taking this approach to analyzing language model behavior more generally can serve as a powerful tool for understanding the intermediate learning phases, overall training dynamics, and the specific generalizations learned by language models.

Country of Origin
🇺🇸 United States

Repos / Data Links

Page Count
14 pages

Category
Computer Science:
Computation and Language