Proceedings of the Fourth Workshop on Online Abuse and Harms 2020
DOI: 10.18653/v1/2020.alw-1.15
|View full text |Cite
|
Sign up to set email alerts
|

Six Attributes of Unhealthy Conversations

Abstract: We present a new dataset of approximately 44000 comments labeled by crowdworkers. Each comment is labelled as either 'healthy' or 'unhealthy', in addition to binary labels for the presence of six potentially 'unhealthy' sub-attributes: (1) hostile; (2) antagonistic, insulting, provocative or trolling; (3) dismissive; (4) condescending or patronising; (5) sarcastic; and/or (6) an unfair generalisation. Each label also has an associated confidence score. We argue that there is a need for datasets which enable re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 30 publications
(24 citation statements)
references
References 25 publications
0
24
0
Order By: Relevance
“…In recent work coming out of a collaboration between Google and researchers from the universities of Oxford and South Carolina, these operative distinctions are becoming more fine-grained, seeking to tackle “subtle forms of toxicity” (Price et al, 2020), which means identifying comments that are “(1) hostile; (2) antagonistic, insulting, provocative or trolling; (3) dismissive; (4) condescending or patronizing; (5) sarcastic; and/or (6) an unfair generalisation” (Price et al, 2020: 1). Whether this work ends up in a production system or not, it signals the deep entanglement between technical design and potentially far-reaching moments of normative prescription in machine moderation.…”
Section: The Fabrics Of Powermentioning
confidence: 99%
“…In recent work coming out of a collaboration between Google and researchers from the universities of Oxford and South Carolina, these operative distinctions are becoming more fine-grained, seeking to tackle “subtle forms of toxicity” (Price et al, 2020), which means identifying comments that are “(1) hostile; (2) antagonistic, insulting, provocative or trolling; (3) dismissive; (4) condescending or patronizing; (5) sarcastic; and/or (6) an unfair generalisation” (Price et al, 2020: 1). Whether this work ends up in a production system or not, it signals the deep entanglement between technical design and potentially far-reaching moments of normative prescription in machine moderation.…”
Section: The Fabrics Of Powermentioning
confidence: 99%
“…NLP-aided approaches to detect abusive behavior online is an active research area (Schmidt and Wiegand, 2017;Mishra et al, 2019;Corazza et al, 2020). Researchers have developed typologies of online abuse (Waseem et al, 2017), constructed datasets annotated with different types of abusive language (Warner and Hirschberg, 2012;Price et al, 2020;Vidgen et al, 2021), and built NLP models to detect them efficiently Mozafari et al, 2019). Researchers have also expanded the focus to more subtle forms of abuse such as condescension and microaggressions (Breitfeller et al, 2019;.…”
Section: Detecting Online Abusementioning
confidence: 99%
“…In another line of work, theoretical foundations are being established, in the form of taxonomies (Banko et al, 2020), definitions (Wiegand et al, 2021;Waseem et al, 2017) and theory (Price et al, 2020;Laaksonen et al, 2020). We are adding to this with definitions based on fieldwork and grounded research, inspired by anthropological and ethnographic work that investigates the societal impact of online hate and extreme speech (Boromisza-Habashi, 2013;Donovan and danah boyd, 2021;Haynes, 2019;Udupa and Pohjonen, 2019;Hervik, 2019).…”
Section: Related Workmentioning
confidence: 99%