Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.221
|View full text |Cite
|
Sign up to set email alerts
|

Features or Spurious Artifacts? Data-centric Baselines for Fair and Robust Hate Speech Detection

Abstract: Warning: this paper contains content that may be offensive or upsetting.Avoiding to rely on dataset artifacts to predict hate speech is at the cornerstone of robust and fair hate speech detection. In this paper we critically analyze lexical biases in hate speech detection via a cross-platform study, disentangling various types of spurious and authentic artifacts and analyzing their impact on out-of-distribution fairness and robustness. We experiment with existing approaches and propose simple yet surprisingly … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(24 citation statements)
references
References 28 publications
0
10
0
Order By: Relevance
“…A thorough analysis on the impact of diverse auxiliary tasks on the performance of our models for PCL detection, and an investigation on the role of uncertainty and disagreement further confirmed the importance of considering annotators' point of view in PCL detection. As future work, we aim to test the presence and assess the impact of spurious lexical biases in the dataset (Ramponi and Tonelli, 2022) and extend our models to other genres, such as social media (Wang and Potts, 2019). We hope this work will encourage future efforts towards annotatorscentric NLP, on PCL detection and other subjective tasks more broadly.…”
Section: Discussionmentioning
confidence: 99%
“…A thorough analysis on the impact of diverse auxiliary tasks on the performance of our models for PCL detection, and an investigation on the role of uncertainty and disagreement further confirmed the importance of considering annotators' point of view in PCL detection. As future work, we aim to test the presence and assess the impact of spurious lexical biases in the dataset (Ramponi and Tonelli, 2022) and extend our models to other genres, such as social media (Wang and Potts, 2019). We hope this work will encourage future efforts towards annotatorscentric NLP, on PCL detection and other subjective tasks more broadly.…”
Section: Discussionmentioning
confidence: 99%
“…• group : the target of the abuse is a protected group or an individual as part of that group. We follow the widely-used definition of protected groups ( Röttger et al, 2021 ; Ramponi & Tonelli, 2022 ; Banko, MacKeen & Ray, 2020 ), namely groups based on characteristics such as religion, ethnicity, race, gender identity, age, sex or sexual orientation, disability, and national origins. The category is related to Davidson et al (2017) ’s “hate speech” definition and focus on protected characteristics.…”
Section: A Taxonomy For Religious Hatementioning
confidence: 99%
“…In this section, we present the protocol we followed for collecting and annotating religious hate speech data in English and Italian. We then provide documentation in the form of data and artifacts statements ( Bender & Friedman, 2018 ; Ramponi & Tonelli, 2022 ), as well as summary statistics and insights about the annotated corpus. While data collection follows the same protocol for both languages, we adopt two different approaches to data annotation.…”
Section: Dataset Creationmentioning
confidence: 99%
See 1 more Smart Citation
“…Our research hypothesis is that the labels given by individuals with a low CRT score may be noisy and biased. To test the hypothesis and support the effectiveness of the dataset, we conducted experiments with varying architectures and metrics that cover human-centered desiderata of hate speech classifiers, such as detection performance, fairness (Ramponi and Tonelli, 2022), and explainability (Mathew et al, 2021).…”
Section: Introductionmentioning
confidence: 99%