2022
DOI: 10.1007/s10639-022-11056-x
|View full text |Cite
|
Sign up to set email alerts
|

The design, construction and evaluation of annotated Arabic cyberbullying corpus

Abstract: Cyberbullying (CB) is classified as one of the severe misconducts on social media. Many CB detection systems have been developed for many natural languages to face this phenomenon. However, Arabic is one of the under-resourced languages suffering from the lack of quality datasets in many computational research areas. This paper discusses the design, construction, and evaluation of a multi-dialect, annotated Arabic Cyberbullying Corpus (ArCybC), a valuable resource for Arabic CB detection and motivation for fut… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 60 publications
0
11
0
Order By: Relevance
“…as (Abainia, 2020;Alsafari et al, 2020;Boucherit & Abainia, 2022;Mubarak et al, 2021;Shannag et al, 2022). Furthermore, some datasets were annotated by their respective authors (Alshehri et al, 2020;Badri et al, 2022;Khairy et al, 2022), while for others, no information was provided regarding the criteria used for selecting the annotators (Alam et al, 2022;De Smedt et al, 2018;Mohdeb et al, 2022;Obeidat et al, 2022;Raïdy & Harmanani, 2023).…”
Section: Notementioning
confidence: 99%
See 3 more Smart Citations
“…as (Abainia, 2020;Alsafari et al, 2020;Boucherit & Abainia, 2022;Mubarak et al, 2021;Shannag et al, 2022). Furthermore, some datasets were annotated by their respective authors (Alshehri et al, 2020;Badri et al, 2022;Khairy et al, 2022), while for others, no information was provided regarding the criteria used for selecting the annotators (Alam et al, 2022;De Smedt et al, 2018;Mohdeb et al, 2022;Obeidat et al, 2022;Raïdy & Harmanani, 2023).…”
Section: Notementioning
confidence: 99%
“…In some works, the crowdworkers are evaluated, without notifying them, by incorporating texts from the pre-annotated sample into each crowdworker task. Less accurate crowdworkers are disqualified based on comparison with the expert labels (Albadi et al, 2018(Albadi et al, , 2022Alhelbawy et al, 2016;Chowdhury et al, 2020;Mubarak, Hassan, & Chowdhury, 2022;Shannag et al, 2022). Another approach is selecting crowdworkers with good reputation scores, which are provided on the crowdsourcing platform (Ousidhoum et al, 2019).…”
Section: Quality and Validationmentioning
confidence: 99%
See 2 more Smart Citations
“…Arabic Cyberbullyig Corpus (ArCybC) [52] is the first publicly available cyberbullying dataset for the Arabic language. Researchers can use it to classify tweets annotated as Cyberbullying (CB), Non-Cyberbullying (Non-CB), Offensive (Off), and Non-Offensive (Non-Off).…”
Section: Dataset Preparationmentioning
confidence: 99%