Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1578
|View full text |Cite
|
Sign up to set email alerts
|

Perturbation Sensitivity Analysis to Detect Unintended Model Biases

Abstract: arXiv:1910.04210v1 [cs.CL]

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 57 publications
(36 citation statements)
references
References 18 publications
0
25
0
2
Order By: Relevance
“…basic negation, agent/object distinction, etc). Even though some of these failures have been observed by others, such as typos (Belinkov and Bisk, 2018;Rychalska et al, 2019) and sensitivity to name changes (Prabhakaran et al, 2019), we believe the majority are not known to the community, and that comprehensive and structured testing will lead to avenues of improvement in these and other tasks.…”
Section: Discussionmentioning
confidence: 86%
See 2 more Smart Citations
“…basic negation, agent/object distinction, etc). Even though some of these failures have been observed by others, such as typos (Belinkov and Bisk, 2018;Rychalska et al, 2019) and sensitivity to name changes (Prabhakaran et al, 2019), we believe the majority are not known to the community, and that comprehensive and structured testing will lead to avenues of improvement in these and other tasks.…”
Section: Discussionmentioning
confidence: 86%
“…There are existing perturbation techniques meant to evaluate specific behavioral capabilities of NLP models such as logical consistency and robustness to noise (Belinkov and Bisk, 2018), name changes (Prabhakaran et al, 2019), or adversaries (Ribeiro et al, 2018). CheckList provides a framework for such techniques to systematically evaluate these alongside a variety of other capabilities.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Following (Garg et al, 2019;Prabhakaran et al, 2019), we use the notion of perturbation, whereby the phrases for referring to people with disabilities, described above, are all inserted into the same slots in sentence templates. We start by first retrieving a set of naturally-occurring sentences that contain the pronouns he or she.…”
Section: Biases In Text Classification Modelsmentioning
confidence: 99%
“…A recent thread of work aims to study how language models recall and leverage information about names and entities. Prabhakaran et al (2019) shows that names can have a measurable effect on the prediction of sentiment analysis systems. Shwartz et al (2020) demonstrates that pre-trained language models implicitly resolve entity ambiguity by grounding names to entities based on the pretraining corpus.…”
Section: Related Workmentioning
confidence: 99%