From Linear Input to Hierarchical Structure: Function Words as Statistical Cues for Language Learning
By: Xiulin Yang, Heidi Getz, Ethan Gotlieb Wilcox
Potential Business Impact:
Helps computers learn language structure faster.
What statistical conditions support learning hierarchical structure from linear input? In this paper, we address this question by focusing on the statistical distribution of function words. Function words have long been argued to play a crucial role in language acquisition due to their distinctive distributional properties, including high frequency, reliable association with syntactic structure, and alignment with phrase boundaries. We use cross-linguistic corpus analysis to first establish that all three properties are present across 186 studied languages. Next, we use a combination of counterfactual language modeling and ablation experiments to show that language variants preserving all three properties are more easily acquired by neural learners, with frequency and structural association contributing more strongly than boundary alignment. Follow-up probing and ablation analyses further reveal that different learning conditions lead to systematically different reliance on function words, indicating that similar performance can arise from distinct internal mechanisms.
Similar Papers
Language Hierarchization Provides the Optimal Solution to Human Working Memory Limits
Computation and Language
Makes language easier to understand by using memory.
Disjoint Processing Mechanisms of Hierarchical and Linear Grammars in Large Language Models
Computation and Language
Computers learn language structure like brains do.
Rethinking the Relationship between the Power Law and Hierarchical Structures
Computation and Language
Language patterns don't prove how brains build sentences.