z-logo
open-access-imgOpen Access
Predictive structure or paradigm size? Investigating the effects of i-complexity and e-complexity on the learnability of morphological systems
Author(s) -
Tamar Johnson,
Kun Gao,
Kenny Smith,
Hugh Rabagliati,
Jennifer Culbertson
Publication year - 2021
Publication title -
journal of language modelling
Language(s) - English
Resource type - Journals
eISSN - 2299-856X
pISSN - 2299-8470
DOI - 10.15398/jlm.v9i1.259
Subject(s) - learnability , range (aeronautics) , computational complexity theory , computer science , descriptive complexity theory , structural complexity , linguistic sequence complexity , function (biology) , structural complexity theory , dimension (graph theory) , computational model , theoretical computer science , artificial intelligence , complexity class , linguistics , mathematics , algorithm , biology , philosophy , materials science , evolutionary biology , composite material , pure mathematics
Research on cross-linguistic differences in morphological paradigms reveals a wide range of variation on many dimensions, including the number of categories expressed, the number of unique forms, and the number of inflectional classes. However, in an influential paper, Ackerman & Malouf (2013) argue that there is one dimension on which languages do not differ widely: in predictive structure. Predictive structure in a paradigm describes the extent to which forms predict each other, called i-complexity. Ackerman & Malouf (2013) show that although languages differ according to measure of surface paradigm complexity, called e-complexity, they tend to have low i-complexity. They conclude that morphological paradigms have evolved under a pressure for low i-complexity, such that even paradigms with very high e-complexity are relatively easy to learn so long as they have low i-complexity. While this would potentially explain why languages are able to maintain large paradigms, recent work by Johnson et al. (submitted) suggests that both neural networks and human learners may actually be more sensitive to e-complexity than i-complexity. Here we will build on this work, reporting a series of experiments under more realistic learning conditions which confirm that indeed, across a range of paradigms that vary in either e- or i-complexity, neural networks (LSTMs) are sensitive to both, but show a larger effect of e-complexity (and other measures associated with size and diversity of forms). In human learners, we fail to find any effect of i-complexity at all. Further, analysis of a large number of randomly generated paradigms show that e- and i-complexity are negatively correlated: paradigms with high e-complexity necessarily show low i-complexity.These findings suggest that the observations made by Ackerman & Malouf (2013) for natural language paradigms may stem from the nature of these measures rather than learning pressures specially attuned to i-complexity.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here