2021
DOI: 10.48550/arxiv.2106.04511
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Designing Toxic Content Classification for a Diversity of Perspectives

Abstract: In this work, we demonstrate how existing classifiers for identifying toxic comments online fail to generalize to the diverse concerns of Internet users. We survey 17,280 participants to understand how user expectations for what constitutes toxic content differ across demographics, beliefs, and personal experiences. We find that groups historically at-risk of harassment-such as people who identify as LGBTQ+ or young adults-are more likely to to flag a random comment drawn from Reddit, Twitter, or 4chan as toxi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(7 citation statements)
references
References 19 publications
(25 reference statements)
0
7
0
Order By: Relevance
“…For the jury composition task, we displayed one of 5 possible comment sets (generated by random samples from our comment toxicity dataset [49] stratified by toxicity severity and labeler disagreement) to exemplify the type of content they would need to moderate on YourPlatform. Participants were then shown a simplified jury composition input form that allowed them to allocate 12-person jury slots using three demographic attributes: (1) gender (Female, Male, Non-binary, Other), ( 2) race (Black or African American, White, Asian, Hispanic, American Indian or Alaska Native, Native Hawaiian or Pacific Islander, Other) and (3) political affiliation (Conservative, Liberal, Independent, Other).…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…For the jury composition task, we displayed one of 5 possible comment sets (generated by random samples from our comment toxicity dataset [49] stratified by toxicity severity and labeler disagreement) to exemplify the type of content they would need to moderate on YourPlatform. Participants were then shown a simplified jury composition input form that allowed them to allocate 12-person jury slots using three demographic attributes: (1) gender (Female, Male, Non-binary, Other), ( 2) race (Black or African American, White, Asian, Hispanic, American Indian or Alaska Native, Native Hawaiian or Pacific Islander, Other) and (3) political affiliation (Conservative, Liberal, Independent, Other).…”
Section: Methodsmentioning
confidence: 99%
“…Moderators of online communities (N=18) were asked to author juries for a comment toxicity classification task. We find that the resulting juries contain 2.9 times the representation of non-White jurors and 31.5 times the representation of non-binary jurors compared to those created implicitly by a large toxicity dataset [49]. This increased diversity in the jury composition changed the algorithm's classifications on 14% of items, reflecting the fact that jury learning captured those individual jurors' views far better than a baseline, state of the art aggregated model (with an MAE of 0.62 versus 1.05).…”
Section: Introductionmentioning
confidence: 92%
See 3 more Smart Citations