2018
DOI: 10.1007/978-3-319-78583-7_3
|View full text |Cite|
|
Sign up to set email alerts
|

Bridging the Gaps: Multi Task Learning for Domain Transfer of Hate Speech Detection

Abstract: Hate speech detection has been the subject of high research attention, due to the scale of content created on social media. In spite of the attention and the sensitive nature of the task, privacy preservation in hate speech detection has remained under-studied. The majority of research has focused on centralised machine learning infrastructures which risk leaking data. In this paper, we show that using federated machine learning can help address privacy the concerns that are inherent to hate speech detection w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
82
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 68 publications
(91 citation statements)
references
References 37 publications
(13 reference statements)
1
82
0
2
Order By: Relevance
“…Table 2 summarized the obtained results for fine-tuning strategies along with the official baselines. We use Waseem and Hovy [22], Davidson et al [3], and Waseem et al [23] as baselines and compare the results with our different fine-tuning strategies using pre-trained BERT base model. The evaluation results are reported on the test dataset and on three different metrics: precision, recall, and weighted-average F1-score.…”
Section: Implementation and Results Analysismentioning
confidence: 99%
See 3 more Smart Citations
“…Table 2 summarized the obtained results for fine-tuning strategies along with the official baselines. We use Waseem and Hovy [22], Davidson et al [3], and Waseem et al [23] as baselines and compare the results with our different fine-tuning strategies using pre-trained BERT base model. The evaluation results are reported on the test dataset and on three different metrics: precision, recall, and weighted-average F1-score.…”
Section: Implementation and Results Analysismentioning
confidence: 99%
“…Zhang et al [25] used a CNN+GRU (Gated Recurrent Unit network) neural network model initialized with pre-trained word2vec embeddings to capture both word/character combinations (e. g., n-grams, phrases) and word/character dependencies (order information). Waseem et al [23] brought a new insight to hate speech and abusive language detection tasks by proposing a multi-task learning framework to deal with datasets across different annotation schemes, labels, or geographic and cultural influences from data sampling. Founta et al [7] built a unified classification model that can efficiently handle different types of abusive language such as cyberbullying, hate, sarcasm, etc.…”
Section: Previous Workmentioning
confidence: 99%
See 2 more Smart Citations
“…We expect that this result occurred for two reasons. First, the dataset contains a large number of cases where AAE is used (Waseem et al, 2018). Second, many of the AAE tweets also use words like "n*gga" and "b*tch", and are thus frequently associated with the hate speech and offensive classes, resulting in "false positive bias" (Dixon et al, 2018).…”
Section: Discussionmentioning
confidence: 99%