2018
DOI: 10.48550/arxiv.1805.12471
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Network Acceptability Judgments

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
95
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 76 publications
(96 citation statements)
references
References 0 publications
0
95
1
Order By: Relevance
“…Datasets. We evaluate our method on the GLUE benchmark (Wang et al, 2018), which consists of eight datasets covering various types of tasks including natural language inference (RTE (Dagan et al, 2005), MNLI (Williams et al, 2018), and QNLI (Rajpurkar et al, 2016)), semantic textual similarity (MRPC (Dolan and Brockett, 2005), STS-B (Cer et al, 2017), and QQP 3 ), linguistic acceptability (CoLA (Warstadt et al, 2018)), and sentiment analysis (SST2 (Socher et al, 2013)).…”
Section: Methodsmentioning
confidence: 99%
“…Datasets. We evaluate our method on the GLUE benchmark (Wang et al, 2018), which consists of eight datasets covering various types of tasks including natural language inference (RTE (Dagan et al, 2005), MNLI (Williams et al, 2018), and QNLI (Rajpurkar et al, 2016)), semantic textual similarity (MRPC (Dolan and Brockett, 2005), STS-B (Cer et al, 2017), and QQP 3 ), linguistic acceptability (CoLA (Warstadt et al, 2018)), and sentiment analysis (SST2 (Socher et al, 2013)).…”
Section: Methodsmentioning
confidence: 99%
“…• BERT fine-tuning on CoLA dataset [39]. We use pretrained BERT from Transformers library [40] (bert-base-uncased) and freeze all layers except the last two linear ones.…”
Section: Numerical Experimentsmentioning
confidence: 99%
“…. , k simultaneously where C is defined in (39). Let E k denote the probabilistic event that this statement holds.…”
Section: B11 Two Lemmasmentioning
confidence: 99%
“…Therefore, in this module, we prune the candidate list and retain only the grammatical ones. Toward this, we train a grammaticality classifier on the corpus of linguistic acceptability (CoLA) (Warstadt et al, 2018), a dataset with 10,657 English sentences labeled as grammatical or ungrammatical from linguistics publications. We select BERT (Devlin et al, 2019) as the classification model, and fine-tune it on the CoLA dataset.…”
Section: Candidate Pruningmentioning
confidence: 99%