2022 ACM Conference on Fairness, Accountability, and Transparency 2022
DOI: 10.1145/3531146.3533216
|View full text |Cite
|
Sign up to set email alerts
|

Assessing Annotator Identity Sensitivity via Item Response Theory: A Case Study in a Hate Speech Corpus

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 34 publications
0
5
0
Order By: Relevance
“…Details on the development of the hate speech measure are documented [ 25 ] in previous research. The training dataset, consisting of 50,000 social media comments sourced from YouTube, Twitter, and Reddit, was labeled by 10,000 US-based Amazon Mechanical Turk [ 26 ] workers on those components of hate speech (the dataset is available at ) (accessed on 21 December 2022) [ 27 ]. Amazon Mechanical Turk [ 26 ] is a crowdsourcing marketplace where tasks can be performed.…”
Section: Methodsmentioning
confidence: 99%
“…Details on the development of the hate speech measure are documented [ 25 ] in previous research. The training dataset, consisting of 50,000 social media comments sourced from YouTube, Twitter, and Reddit, was labeled by 10,000 US-based Amazon Mechanical Turk [ 26 ] workers on those components of hate speech (the dataset is available at ) (accessed on 21 December 2022) [ 27 ]. Amazon Mechanical Turk [ 26 ] is a crowdsourcing marketplace where tasks can be performed.…”
Section: Methodsmentioning
confidence: 99%
“…Similarly, disagreement convolution (Gordon et al 2021) incorporates disagreement fully into the evaluation pipeline, showing that on natural language tasks related to toxicity and misinformation the performance of traditional ML models is often overstated. Furthermore, Basile (2021), Rizos and Schuller (2020) and Sachdeva et al (2022b) showed an additional advantage of strong perspectivism in supervised learning, namely its potential impact on the interpretability and fairness of the models. In experiments on real data ( annotated corpora of hate speech), Basile (2021) showed how individual labels can be used to cluster the raters by affinity, leading to the emergence of patterns that helps identifying sociodemographic aspects of the raters themselves, which are in principle opaque, especially in a crowdsourcing scenario, while Sachdeva et al (2022b) showed how the same information could be useful to assess annotator identity sensitivity and thus identify biases in annotation patterns; also Rizos and Schuller (2020) described how a similar approach could be used to detect biases in the data and labels provided by raters.…”
Section: Review Of Perspectivist Approaches In Aimentioning
confidence: 99%
“…Furthermore, Basile (2021), Rizos and Schuller (2020) and Sachdeva et al (2022b) showed an additional advantage of strong perspectivism in supervised learning, namely its potential impact on the interpretability and fairness of the models. In experiments on real data ( annotated corpora of hate speech), Basile (2021) showed how individual labels can be used to cluster the raters by affinity, leading to the emergence of patterns that helps identifying sociodemographic aspects of the raters themselves, which are in principle opaque, especially in a crowdsourcing scenario, while Sachdeva et al (2022b) showed how the same information could be useful to assess annotator identity sensitivity and thus identify biases in annotation patterns; also Rizos and Schuller (2020) described how a similar approach could be used to detect biases in the data and labels provided by raters. Far from being an exhausted topic, the discussion over perspectivism has recently been fostered, among other venues, at international workshops such as Investigating and Mitigating Biases in Crowdsourced Data 5 (ACM CSCW 2021) and the 1st Workshop on Perspectivist Approaches to NLP 6 .…”
Section: Review Of Perspectivist Approaches In Aimentioning
confidence: 99%
“…Biased annotators could inadvertently induce their implicit prejudices into the labels (Sap et al, 2019;Davidson et al, 2019;Sachdeva et al, 2022). If machine learning models are trained on such biased data, they potentially amplify these biases and generate harm by inference (Shah et al, 2020).…”
Section: Cognitive Reflection Testmentioning
confidence: 99%
“…(3) Biased and noisy labels: Manual annotation is a predominant approach for constructing labeled data. However, the labeling process could reflect inherent biases of human annotators, and it is hard to control the label quality (Sap et al, 2019;Sachdeva et al, 2022). Training a model on such a noisy and biased dataset can amplify biases by inference (Shah et al, 2020).…”
Section: Introductionmentioning
confidence: 99%