Modelling Child Learning and Parsing of Long-range Syntactic Dependencies
By: Louis Mahon, Mark Johnson, Mark Steedman
Potential Business Impact:
Teaches computers how kids learn language.
This work develops a probabilistic child language acquisition model to learn a range of linguistic phenonmena, most notably long-range syntactic dependencies of the sort found in object wh-questions, among other constructions. The model is trained on a corpus of real child-directed speech, where each utterance is paired with a logical form as a meaning representation. It then learns both word meanings and language-specific syntax simultaneously. After training, the model can deduce the correct parse tree and word meanings for a given utterance-meaning pair, and can infer the meaning if given only the utterance. The successful modelling of long-range dependencies is theoretically important because it exploits aspects of the model that are, in general, trans-context-free.
Similar Papers
Step-by-step Instructions and a Simple Tabular Output Format Improve the Dependency Parsing Accuracy of LLMs
Computation and Language
Helps computers understand sentences perfectly.
Leveraging Large Language Models for Robot-Assisted Learning of Morphological Structures in Preschool Children with Language Vulnerabilities
Robotics
Robot helps kids with talking problems learn words.
Looking beyond the next token
Machine Learning (CS)
Teaches computers to plan and write stories better.