Depth-Bounded Statistical PCFG Induction as a Model of Human Grammar Acquisition
Author(s) -
Lifeng Jin,
Lane Schwartz,
Finale DoshiVelez,
Timothy M. Miller,
William Schuler
Publication year - 2021
Publication title -
computational linguistics
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.314
H-Index - 98
eISSN - 1530-9312
pISSN - 0891-2017
DOI - 10.1162/coli_a_00399
Subject(s) - rule based machine translation , computer science , grammar , embedding , simple (philosophy) , grammar induction , natural language processing , bounded function , l attributed grammar , set (abstract data type) , domain (mathematical analysis) , artificial intelligence , context sensitive grammar , tree adjoining grammar , theoretical computer science , programming language , linguistics , mathematics , context free grammar , mathematical analysis , philosophy , epistemology
This article describes a simple PCFG induction model with a fixed category domain that predicts a large majority of attested constituent boundaries, and predicts labels consistent with nearly half of attested constituent labels on a standard evaluation data set of child-directed speech. The article then explores the idea that the difference between simple grammars exhibited by child learners and fully recursive grammars exhibited by adult learners may be an effect of increasing working memory capacity, where the shallow grammars are constrained images of the recursive grammars. An implementation of these memory bounds as limits on center embedding in a depth-specific transform of a recursive grammar yields a significant improvement over an equivalent but unbounded baseline, suggesting that this arrangement may indeed confer a learning advantage.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom