Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1356
|View full text |Cite
|
Sign up to set email alerts
|

Studying the Inductive Biases of

Abstract: How do typological properties such as word order and morphological case marking affect the ability of neural sequence models to acquire the syntax of a language? Crosslinguistic comparisons of RNNs' syntactic performance (e.g., on subject-verb agreement prediction) are complicated by the fact that any two languages differ in multiple typological properties, as well as by differences in training corpus. We propose a paradigm that addresses these issues: we create synthetic versions of English, which differ from… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
16
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 29 publications
(19 citation statements)
references
References 24 publications
(26 reference statements)
2
16
1
Order By: Relevance
“…Because we have the same amount of training data per-language in the same domain, this could point to the importance of having explicit cues to lin-guistic structure such that models can learn that structure. While more language varieties need to be evaluated to determine whether this trend is robust, we note that this finding is consistent with that of Ravfogel et al (2019), who compared English to a synthetic variety of English augmented with case markers and found that the addition of case markers increased LSTM agreement prediction accuracy.…”
Section: Morphological Complexity Vs Accuracysupporting
confidence: 83%
See 1 more Smart Citation
“…Because we have the same amount of training data per-language in the same domain, this could point to the importance of having explicit cues to lin-guistic structure such that models can learn that structure. While more language varieties need to be evaluated to determine whether this trend is robust, we note that this finding is consistent with that of Ravfogel et al (2019), who compared English to a synthetic variety of English augmented with case markers and found that the addition of case markers increased LSTM agreement prediction accuracy.…”
Section: Morphological Complexity Vs Accuracysupporting
confidence: 83%
“…Dhar and Bisazza (2018) trained a multilingual LM on a concatenated French and Italian corpus, and tested whether grammatical abilities transfer across languages. Ravfogel et al (2018) reported an in-depth analysis of LSTM LM performance on agreement prediction in Basque, and Ravfogel et al (2019) investigated the effect of different syntactic properties of a language on RNNs' agreement prediction accuracy by creating synthetic variants of English. Finally, grammatical evaluation has been proposed for machine translation systems for languages such as German and French (Sennrich, 2017;Isabelle et al, 2017).…”
Section: Grammatical Evaluation Beyond Englishmentioning
confidence: 99%
“…This result indicates some architectural limitations of LSTM-LMs in handling object RCs robustly at a near perfect level. Answering why the accuracy does not reach (almost) 100%, perhaps with other empirical properties or inductive biases (Khandelwal et al, 2018;Ravfogel et al, 2019) is future work.…”
Section: Limitations Of Lstm-lmsmentioning
confidence: 99%
“…However, Kuncoro et al (2018) have also shown that although sequential LSTMs can learn syntactic information, a recursive neural network that explicitly models hierarchy (the Recurrent Neural Network Grammar model from Dyer et al [2015]) is better at this: It performs better on the number agreement task from Linzen, Dupoux, and Goldberg (2016). In addition, Ravfogel, Goldberg, and Tyers (2018) and Ravfogel, Goldberg, and Linzen (2019) have cast some doubts on the results by Linzen, Dupoux, and Goldberg (2016) and Gulordava et al (2018) by looking at Basque and synthetic languages with different word orders, respectively, in the two studies.…”
Section: Recursive Vs Recurrent Neural Networkmentioning
confidence: 99%