Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing 2022
DOI: 10.18653/v1/2022.emnlp-main.744
|View full text |Cite
|
Sign up to set email alerts
|

KOLD: Korean Offensive Language Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…Compared to the reported performance on other datasets, there was a broad range of values. For instance, the KOLD dataset [37] reported precision and recall rates of 50.8 and 47.8, respectively. In contrast, the best model in the "SemEval-2021 Task 5" toxic-spandetection competition, Ref.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Compared to the reported performance on other datasets, there was a broad range of values. For instance, the KOLD dataset [37] reported precision and recall rates of 50.8 and 47.8, respectively. In contrast, the best model in the "SemEval-2021 Task 5" toxic-spandetection competition, Ref.…”
Section: Discussionmentioning
confidence: 99%
“…They achieved a mean F1-score of 77.16% with PhoBERT large and 77.70% using XLM-RoBERTa large . Another recently released dataset is the Korean Offensive Language Dataset (KOLD) [37], which offers a hierarchical taxonomy and annotations at the span level for identifying toxic content. In addition to the taxonomy, similar to OLIF [15], Jeong et al [37] proposed the labeling of the target group, similar to the approach of HateXplain [29].…”
Section: Offensive and Toxic Spans' Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…Many studies have presented benchmark models and provide a number of moral frameworks for reducing risk prior to product release. For example, studies aim to prevent physical harm (Levy et al, 2022) or hate speech by establishing benchmarks (Jeong et al, 2022), consistently contributing to practical design references. However, the paper, so-called "the Salmon paper" criticizes some of existing benchmark datasets that are designed to measure stereotyping (Su et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Offensive language can vary greatly depending on cultural backgrounds. While most multilingual OLD datasets are constructed by filtering a predefined list of offensive words (Zampieri et al, 2019;Sigurbergsson and Derczynski, 2020;Jeong et al, 2022;Deng et al, 2022), certain offensive words are culturally specific. For example, OLD models trained on American cultural contexts may struggle to effectively detect offensive words like "m*adarchod" and "pr*sstitute" in Indian texts (Ghosh et al, 2021;Santy et al, 2023).…”
Section: Introductionmentioning
confidence: 99%