Learning Syntactic Structures from String Input
Ethan Wilcox, Jon Gauthier, Jennifer Hu, Peng Qian, Roger Levy
February 2022
 

This chapter addresses a series of interrelated questions about the origin of syntactic structures: How do language learners generalize from the linguistic stimulus with which they are presented? To what extent does linguistic cognition recruit domain-general (i.e. not language-specific) processes and representations? And to what extent are rules and generalizations about linguistic structure separate from rules and generalizations about linguistic meaning? We address these questions by asking what syntactic generalizations can be acquired by a domain-general learner from string input alone. The learning algorithm we deploy is a neural-network based Language Model (GPT-2; Radford et al., 2019), which has been trained to provide probability distributions over strings of text. We assess its linguistic capabilities by treating it like a human subject in a psycholinguistics experiment, and inspect behavior in controlled, factorized tests that are designed to reveal the learning outcomes for one particular syntactic generalization. The tests presented in this chapter focus on a variety of syntactic phenomena in two broad categories: rules about the structure of the sentence and rules about the relationships between smaller lexical units, including scope and binding. Results indicate that our target model has learned many subtle syntactic generalizations, yet it still falls short of humanlike grammatical competence in some areas, notably for cases of parasitic gaps (e.g. “I know what you burned ___ after reading ___ yesterday”). We discuss the implications of these results under three interpretive frameworks, which view the model as (a) a counter-argument against claims of linguistic innateness, (b) a positive example of syntactic emergentism, and (c) a fully-articulated model of grammatical competence.
Format: [ pdf ]
Reference: lingbuzz/007271
(please use that when you cite this article)
Published in: Algebraic Structures in Natural Language
keywords: syntax, learnability, deep learning, psycholinguistics, syntax
Downloaded:365 times

 

[ edit this article | back to article list ]