2020
DOI: 10.48550/arxiv.2004.12764
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

"Call me sexist, but...": Revisiting Sexism Detection Using Psychological Scales and Adversarial Samples

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…The latter issue is particularly relevant to hate speech detection since current hate speech datasets vary in data source, sampling strategy and annotation process (Vidgen and Derczynski, 2020;Poletto et al, 2020), and are known to exhibit annotator biases (Waseem, 2016;Waseem et al, 2018;Sap et al, 2019) as well as topic and author biases (Wiegand et al, 2019;Nejadgholi and Kiritchenko, 2020). Correspondingly, models trained on such datasets have been shown to be overly sensitive to lexical features such as group identifiers (Park et al, 2018;Dixon et al, 2018;Kennedy et al, 2020), and to generalise poorly to other datasets (Nejadgholi and Kiritchenko, 2020;Samory et al, 2020). Therefore, held-out performance on current hate speech datasets is an incomplete and potentially misleading measure of model quality.…”
Section: Introductionmentioning
confidence: 99%
“…The latter issue is particularly relevant to hate speech detection since current hate speech datasets vary in data source, sampling strategy and annotation process (Vidgen and Derczynski, 2020;Poletto et al, 2020), and are known to exhibit annotator biases (Waseem, 2016;Waseem et al, 2018;Sap et al, 2019) as well as topic and author biases (Wiegand et al, 2019;Nejadgholi and Kiritchenko, 2020). Correspondingly, models trained on such datasets have been shown to be overly sensitive to lexical features such as group identifiers (Park et al, 2018;Dixon et al, 2018;Kennedy et al, 2020), and to generalise poorly to other datasets (Nejadgholi and Kiritchenko, 2020;Samory et al, 2020). Therefore, held-out performance on current hate speech datasets is an incomplete and potentially misleading measure of model quality.…”
Section: Introductionmentioning
confidence: 99%