Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.310
|View full text |Cite
|
Sign up to set email alerts
|

On the Robustness of Language Encoders against Grammatical Errors

Abstract: We conduct a thorough study to diagnose the behaviors of pre-trained language encoders (ELMo, BERT, and RoBERTa) when confronted with natural grammatical errors. Specifically, we collect real grammatical errors from non-native speakers and conduct adversarial attacks to simulate these errors on clean text data. We use this approach to facilitate debugging models on downstream applications. Results confirm that the performance of all tested models is affected but the degree of impact varies. To interpret model … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(9 citation statements)
references
References 30 publications
0
5
0
Order By: Relevance
“…This fixed adversarial test set will then be used to evaluate all the new models. This evaluation setup has been adopted in (Ren et al, 2019;Tan et al, 2020;Yin et al, 2020;Wang et al, 2020b;Zou et al, 2020;Wang et al, 2021, inter alia. ).…”
Section: Robustness Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…This fixed adversarial test set will then be used to evaluate all the new models. This evaluation setup has been adopted in (Ren et al, 2019;Tan et al, 2020;Yin et al, 2020;Wang et al, 2020b;Zou et al, 2020;Wang et al, 2021, inter alia. ).…”
Section: Robustness Evaluationmentioning
confidence: 99%
“…However, they are limited in practice as they are not generally applicable to other types of attacks. The other type of defense is Adversarial Data Augmentation (ADA), which augments the training set by the adversarial examples and is widely used in the training (finetuning) process to enhance model robustness (Alzantot et al;Ren et al, 2019;Jin et al, 2020;Li et al, 2020;Tan et al, 2020;Yin et al, 2020;Zheng et al, 2020;Zou et al, 2020;Wang et al, 2020b). ADA is generally applicable to any type of adversarial attacks but is not very effective in improving model performance under attacks.…”
Section: Introductionmentioning
confidence: 99%
“…The ability of these models to detect and locate grammatical errors was studied in [ 23 ]. Error examples were taken from the NUCLE (NUS Corpus of Learner English) dataset consisting of pairs of erroneous and correct sentences.…”
Section: Related Workmentioning
confidence: 99%
“…The vulnerability of modern neural networks towards human imperceptible input variations has been studied for a while since (Szegedy et al, 2013), primarily in the computer vision community (e.g., (Goodfellow et al, 2015)), later extended to the NLP community (e.g., (Ebrahimi et al, 2017;Liang et al, 2017;Yin et al, 2020;Jones et al, 2020;Jia et al, 2019;Liu et al, 2019;Pruthi et al, 2019)). Recent studies suggest that the fragility of neural networks roots in that the data has multiple signals that can reduce the empirical risk, and when a model is forced to reduce the training error, it picks up whatever information that diminish the empirical loss, ignoring whether the learnt knowledge aligns with human perception or not (Wang et al, 2019b), connecting the adversarial robustness problems and the bias in data problems that has been studied for a while (e.g., (Wang et al, 2016;Goyal et al, 2017;Kaushik and Lipton, 2018;Wang et al, 2019a)).…”
Section: Related Workmentioning
confidence: 99%