2019
DOI: 10.1162/tacl_a_00290
|View full text |Cite
|
Sign up to set email alerts
|

Neural Network Acceptability Judgments

Abstract: This paper investigates the ability of artificial neural networks to judge the grammatical acceptability of a sentence. Machine learning research of this kind is well placed to answer important open questions about the role of prior linguistic bias in language acquisition by providing a test for the Poverty of the Stimulus Argument. In service of this goal, we introduce the Corpus of Linguistic Acceptability (CoLA), a set of 10,657 English sentences labeled as grammatical or ungrammatical from published lingui… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
371
1
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 614 publications
(446 citation statements)
references
References 48 publications
0
371
1
1
Order By: Relevance
“…nine natural language understanding (NLU) tasks. As shown in Table 1, it includes question answering (Rajpurkar et al, 2016), linguistic acceptability (Warstadt et al, 2018), sentiment analysis (Socher et al, 2013), text similarity (Cer et al, 2017), paraphrase detection (Dolan and Brockett, 2005), and natural language inference (NLI) Bar-Haim et al, 2006;Giampiccolo et al, 2007;Bentivogli et al, 2009;Levesque et al, 2012;Williams et al, 2018). The diversity of the tasks makes GLUE very suitable for evaluating the generalization and robustness of NLU models.…”
Section: Modelmentioning
confidence: 99%
“…nine natural language understanding (NLU) tasks. As shown in Table 1, it includes question answering (Rajpurkar et al, 2016), linguistic acceptability (Warstadt et al, 2018), sentiment analysis (Socher et al, 2013), text similarity (Cer et al, 2017), paraphrase detection (Dolan and Brockett, 2005), and natural language inference (NLI) Bar-Haim et al, 2006;Giampiccolo et al, 2007;Bentivogli et al, 2009;Levesque et al, 2012;Williams et al, 2018). The diversity of the tasks makes GLUE very suitable for evaluating the generalization and robustness of NLU models.…”
Section: Modelmentioning
confidence: 99%
“…The General Language Understanding Evaluation (GLUE) benchmark ) is a collection of diverse natural language understanding tasks (Warstadt et al, 2018;Socher et al, 2013;Dolan and Brockett, 2005;Agirre et al, 2007;Williams et al, 2018;Rajpurkar et al, 2016;Dagan et al, 2006;Levesque et al, 2011), which is the main benchmark used in Devlin et al (2019).…”
Section: Gluementioning
confidence: 99%
“…The Corpus of Linguistic Acceptability (CoLA) is a binary classification task. The goal of this task is to predict whether an English sentence is linguistically acceptable or not (Warstadt et al, 2018). Table 8 presents the accuracy scores of BERT and DISP on the CoLA dataset with one adversarial attack of each type.…”
Section: Resultsmentioning
confidence: 99%