Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021) 2021
DOI: 10.18653/v1/2021.semeval-1.27
|View full text |Cite
|
Sign up to set email alerts
|

NLRG at SemEval-2021 Task 5: Toxic Spans Detection Leveraging BERT-based Token Classification and Span Prediction Techniques

Abstract: Toxicity detection of text has been a popular NLP task in the recent years. In SemEval-2021 Task-5 Toxic Spans Detection, the focus is on detecting toxic spans within English passages. Most state-of-the-art span detection approaches employ various techniques, each of which can be broadly classified into Token Classification or Span Prediction approaches. In our paper, we explore simple versions of both of these approaches and their performance on the task. Specifically, we use BERT-based models -BERT, RoBERTa,… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(8 citation statements)
references
References 20 publications
0
6
0
Order By: Relevance
“…Since SE tasks require highly nuanced semantic understanding, most solutions leveraged large language models pre-trained using transformers, including BERT (Devlin et al 2019) and other types of transformers (Morio et al 2020;Chhablani et al 2021). These models are pre-trained on billions of words of English text data and can be easily fine-tuned to adapt to new tasks.…”
Section: Span Extractionmentioning
confidence: 99%
See 3 more Smart Citations
“…Since SE tasks require highly nuanced semantic understanding, most solutions leveraged large language models pre-trained using transformers, including BERT (Devlin et al 2019) and other types of transformers (Morio et al 2020;Chhablani et al 2021). These models are pre-trained on billions of words of English text data and can be easily fine-tuned to adapt to new tasks.…”
Section: Span Extractionmentioning
confidence: 99%
“…To the best of our knowledge, this is the first work on extracting hate speech spans. In Section 2.3, we described the propaganda (Da San Martino et al 2020) and toxic (Chhablani et al 2021) SE tasks.…”
Section: Comparison With Other Workmentioning
confidence: 99%
See 2 more Smart Citations
“…It can also be detected by ferreting out offensive and toxic spans in the texts. A toxic span detecting system was developed by leveraging token classification and span prediction techniques that are based on bidirectional encoder representations from transformers (BERT) [36]. Multi-lingual detection of offensive spans (MUDES) [37] was developed to detect offensive spans in texts.…”
Section: Offensive Language Identificationmentioning
confidence: 99%