Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1004
|View full text |Cite
|
Sign up to set email alerts
|

Neural language models as psycholinguistic subjects: Representations of syntactic state

Abstract: We deploy the methods of controlled psycholinguistic experimentation to shed light on the extent to which the behavior of neural network language models reflects incremental representations of syntactic state. To do so, we examine model behavior on artificial sentences containing a variety of syntactically complex structures. We test four models: two publicly available LSTM sequence models of English (Jozefowicz et al., 2016;Gulordava et al., 2018) trained on large datasets; an RNNG (Dyer et al., 2016) trained… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

5
105
0
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
3
1

Relationship

1
9

Authors

Journals

citations
Cited by 127 publications
(131 citation statements)
references
References 42 publications
5
105
0
1
Order By: Relevance
“…Some previous work does use targeted tests to examine specific capacities of LMs-often inspired by psycholinguistic methods. However, the majority of this work has focused on syntactic capabilities of LMs (Linzen et al, 2016;Gulordava et al, 2018;Marvin and Linzen, 2018;Wilcox et al, 2018;Futrell et al, 2019). Relevant to our case study here, using several of these tests Goldberg (2019) shows the BERT model to perform impressively on such syntactic diagnostics.…”
Section: Introductionmentioning
confidence: 68%
“…Some previous work does use targeted tests to examine specific capacities of LMs-often inspired by psycholinguistic methods. However, the majority of this work has focused on syntactic capabilities of LMs (Linzen et al, 2016;Gulordava et al, 2018;Marvin and Linzen, 2018;Wilcox et al, 2018;Futrell et al, 2019). Relevant to our case study here, using several of these tests Goldberg (2019) shows the BERT model to perform impressively on such syntactic diagnostics.…”
Section: Introductionmentioning
confidence: 68%
“…Answering this question is important both for technical outcomes-models with explicit hierarchical structure show performance gains, at least when training on relatively small datasets (Choe and Charniak, 2016;-and for the scientific aim of understanding what biases, learning objectives and training regimes led to humanlike linguistic knowledge. Previous work has approached this question by either examining models' internal state (Weiss et al, 2018;Mareček and Rosa, 2018) or by studying model behavior (Elman, 1991;Linzen et al, 2016;Futrell et al, 2019;McCoy et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…Although LSTMs and GRUs have already been applied to account for human language performance measures (Futrell et al, 2019;Goodkind & Bicknell, 2018;Gulordava, Bojanowski, Grave, Linzen, & Baroni, 2018;Hahn & Keller, 2016;McCoy, Frank, & Linzen, 2018;Sakaguchi, Duh, Post, & Durme, 2017;Van Schijndel & Linzen, 2018a, 2018b, the question remains whether they form more accurate cognitive processing models than traditional SRNs, beyond what might be expected from their stronger language modeling abilities.…”
Section: Introductionmentioning
confidence: 99%