Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.309
|View full text |Cite
|
Sign up to set email alerts
|

An Analysis of the Utility of Explicit Negative Examples to Improve the Syntactic Abilities of Neural Language Models

Abstract: We explore the utilities of explicit negative examples in training neural language models. Negative examples here are incorrect words in a sentence, such as barks in *The dogs barks. Neural language models are commonly trained only on positive examples, a set of sentences in the training data, but recent studies suggest that the models trained in this way are not capable of robustly handling complex syntactic constructions, such as long-distance agreement. In this paper, we first demonstrate that appropriately… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 14 publications
(18 citation statements)
references
References 21 publications
(41 reference statements)
0
16
1
Order By: Relevance
“…This suggests that all monolingual models learned the basic facts of agreement, and were able to apply them to the vocabulary items in our materials. At the other end of the spectrum, performance was only slightly higher than chance in the Across an Object Relative Clause condition for all languages except German, suggesting that LSTMs tend to struggle with center embedding-that is, when a subject-verb dependency is nested within another dependency of the same kind (Marvin and Linzen, 2018;Noji and Takamura, 2020).…”
Section: Lstmsmentioning
confidence: 99%
“…This suggests that all monolingual models learned the basic facts of agreement, and were able to apply them to the vocabulary items in our materials. At the other end of the spectrum, performance was only slightly higher than chance in the Across an Object Relative Clause condition for all languages except German, suggesting that LSTMs tend to struggle with center embedding-that is, when a subject-verb dependency is nested within another dependency of the same kind (Marvin and Linzen, 2018;Noji and Takamura, 2020).…”
Section: Lstmsmentioning
confidence: 99%
“…Our work is in this spirit for negations. Noji and Takamura (2020) propose taking advantage of negative examples and unlikelihood in the training of language models to increase their syntactic abilities. Similarly, Min et al (2020) show the effectiveness of syntactic data augmentation in the case of robustness in NLI.…”
Section: Related Workmentioning
confidence: 99%
“…where δ is the margin between the log-likelihoods of x and x * . This was originally proposed for analyzing the syntactic abilities of language models (Noji and Takamura, 2020). This loss is useful for developing better language models.…”
Section: Sentence-level Margin Loss (Sent)mentioning
confidence: 99%
“…Huang et al (2018) introduced a margin loss that estimates the quality of each beam-searched candidate by comparing it with the reference sentence. More recently, Noji and Takamura (2020) showed negative examples help to improve the syntactic ability of neural language models. They created negative instances from original instances by injecting a grammatical error and used them to calculate a margin loss that will be added to the cross-entropy loss.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation