Proceedings of the Ninth International Workshop on Natural Language Processing for Social Media 2021
DOI: 10.18653/v1/2021.socialnlp-1.8
|View full text |Cite
|
Sign up to set email alerts
|

Understanding and Interpreting the Impact of User Context in Hate Speech Detection

Abstract: As hate speech spreads on social media and online communities, research continues to work on its automatic detection. Recently, recognition performance has been increasing thanks to advances in deep learning and the integration of user features. This work investigates the effects that such features can have on a detection model. Unlike previous research, we show that simple performance comparison does not expose the full impact of including contextualand user information. By leveraging explainability technique… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(9 citation statements)
references
References 25 publications
0
8
0
Order By: Relevance
“…Although our methodology excludes works that are published in venues other than ACL conferences and workshops, we believe that it gives a good indication of the status of XAI in fairness and bias research in NLP. Mosca et al (2021) use SHAP to demonstrate that adding user features to a hate speech detection model reduces biases that are due to spurious correlations in text, but introduces other biases based on user information. Wich et al (2020) also apply SHAP to two example inputs in order to illustrate the political bias of a hate speech model.…”
Section: Applications Of Xai In Fair Nlpmentioning
confidence: 99%
“…Although our methodology excludes works that are published in venues other than ACL conferences and workshops, we believe that it gives a good indication of the status of XAI in fairness and bias research in NLP. Mosca et al (2021) use SHAP to demonstrate that adding user features to a hate speech detection model reduces biases that are due to spurious correlations in text, but introduces other biases based on user information. Wich et al (2020) also apply SHAP to two example inputs in order to illustrate the political bias of a hate speech model.…”
Section: Applications Of Xai In Fair Nlpmentioning
confidence: 99%
“…(4) As non-grouped random raters representing diversity to understand impact on annotations [26,44] (5) As identity-trained classifier-evaluation where classifiers trained on different pools of raters based on markers like gender, first language, age and education are evaluated as to how they perform in terms of precision and recall [4]. (6) As effects of comment authors' demographic differences and not the annotators' demographics as we study [20,25,27,43], or creating new data sets based on political sub-communities of authors [40]. (7) As focusing on model-robustness-evaluation to detect identity markers like race and gender correctly [36], multi-lingual hate speech classification [5], and model-fairness-evaluation [17,19].…”
Section: Related Workmentioning
confidence: 99%
“…Wang (2018) proposes an adaptation of explainability techniques for computer vision to visualize and understand the CNN-GRU classifier for hate speech (Zhang et al, 2018). Mosca et al (2021) (Bach et al, 2015), and a selfexplanatory model (LSTM with an attention mechanism). SHAP explainer is applied (Wich et al, 2020) to investigate the impact of political bias on hate speech classification.…”
Section: Related Workmentioning
confidence: 99%