Proceedings of the Fourth Workshop on Online Abuse and Harms 2020
DOI: 10.18653/v1/2020.alw-1.7
|View full text |Cite
|
Sign up to set email alerts
|

Impact of Politically Biased Data on Hate Speech Classification

Abstract: One challenge that social media platforms are facing nowadays is hate speech. Hence, automatic hate speech detection has been increasingly researched in recent years -in particular with the rise of deep learning. A problem of these models is their vulnerability to undesirable bias in training data. We investigate the impact of political bias on hate speech classification by constructing three politicallybiased data sets (left-wing, right-wing, politically neutral) and compare the performance of classifiers tra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
31
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
3

Relationship

3
7

Authors

Journals

citations
Cited by 41 publications
(31 citation statements)
references
References 29 publications
0
31
0
Order By: Relevance
“…Since unintended bias in hate speech datasets can impair the model's performance (Waseem, 2016) and fairness (Vidgen et al, 2019a;Dixon et al, 2018), a lot of recent work has been done to investigate this phenomenon (Wiegand et al, 2019;Kim et al, 2020). Some work examined racial bias (Sap et al, 2019;Davidson et al, 2019;Xia et al, 2020), others explored gender bias (Gold and Zesch, 2018), aggregation bias (Balayn et al, 2018) and political bias (Wich et al, 2020b). The type of bias we are examining in this study is the annotator bias.…”
Section: Related Workmentioning
confidence: 97%
“…Since unintended bias in hate speech datasets can impair the model's performance (Waseem, 2016) and fairness (Vidgen et al, 2019a;Dixon et al, 2018), a lot of recent work has been done to investigate this phenomenon (Wiegand et al, 2019;Kim et al, 2020). Some work examined racial bias (Sap et al, 2019;Davidson et al, 2019;Xia et al, 2020), others explored gender bias (Gold and Zesch, 2018), aggregation bias (Balayn et al, 2018) and political bias (Wich et al, 2020b). The type of bias we are examining in this study is the annotator bias.…”
Section: Related Workmentioning
confidence: 97%
“…The authors leverage attention scores to quantify the relevance of different input features. Wich et al (2020) applies posthoc explainability on a custom dataset in German to expose and estimate the impact of political bias on hate speech classifiers. More in detail, left-and right-wing political bias within the training data is visualized via DeepSHAP-based explanations (Lundberg and Lee, 2017).…”
Section: Explainability For Recognition Modelsmentioning
confidence: 99%
“…[9] reported problems with the association of minority group language with hate in their data, while [47] have done work on the influence of different biases in the sampling of popular abusive language datasets (e.g., topic and author bias). [46] analyzed how political bias influence hate speech classification models. [37] proposed social bias frames, which is a formalism that "aims to model the pragmatic frames in which people project social biases and stereotypes onto others" [37, p. 1].…”
Section: Related Workmentioning
confidence: 99%