2018
DOI: 10.1007/978-3-030-00350-0_40
|View full text |Cite
|
Sign up to set email alerts
|

Fighting Adversarial Attacks on Online Abusive Language Moderation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 10 publications
0
5
0
Order By: Relevance
“…Harmful [91,156,245] and toxic languages [113,166,224] are not defined precisely in literature, except for toxicity in video games. However, toxic speech is described in a crowdsourcing task to collect a dataset of toxic comments [292] 7 and insists on the type of language used, which motivated our characterization choice.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Harmful [91,156,245] and toxic languages [113,166,224] are not defined precisely in literature, except for toxicity in video games. However, toxic speech is described in a crowdsourcing task to collect a dataset of toxic comments [292] 7 and insists on the type of language used, which motivated our characterization choice.…”
Section: Methodsmentioning
confidence: 99%
“…Park et al [193] use the False Positive and False Negative Equality Differences to quantify gender biases. Some publications assess the time taken to train the models or the time to detect the OCL [165,215,224,300]. Some papers further study the performance of the models by investigating in more detail the types of sentences usually missclassified.…”
Section: Evaluation Metricmentioning
confidence: 99%
See 1 more Smart Citation
“…These efforts, however, are not scalable enough to curtail the rapid growth of toxic content on online platforms ( Davidson et al, 2017 ). There is also the psychological distress associated with exposing human moderators to firsthand accounts of toxicity ( Rodriguez & Rojas-Galeano, 2018 ). These challenges call for developing effective automatic or semiautomatic solutions to detect toxicity from a large stream of content on online platforms.…”
Section: Introductionmentioning
confidence: 99%
“…Users of social media platforms have various reasons for spreading harmful content, like personal or social gain ( Squicciarini, Dupont & Chen, 2014 ). Studies show that publishing toxic content ( i.e., toxic behavior) is contagious ( Tsikerdekis & Zeadally, 2014 ; Rodriguez & Rojas-Galeano, 2018 ); the malicious behavior of users can influence non-malicious users and leads them to misbehave, which affects the overall well-being of online communities. As an example of toxic behavior ( Alfonso & Morris, 2013 ), one Reddit user named Violentacrez created several communities on controversial topics such as gore, and his followers mimicked this behavior by creating communities with highly offensive content as well.…”
Section: Introductionmentioning
confidence: 99%