Proceedings of the Third Workshop on Abusive Language Online 2019
DOI: 10.18653/v1/w19-3502
|View full text |Cite
|
Sign up to set email alerts
|

Exploring Deep Multimodal Fusion of Text and Photo for Hate Speech Classification

Abstract: Interactions among users on social network platforms are usually positive, constructive and insightful. However, sometimes people also get exposed to objectionable content such as hate speech, bullying, and verbal abuse etc. Most social platforms have explicit policy against hate speech because it creates an environment of intimidation and exclusion, and in some cases may promote real-world violence. As users' interactions on today's social networks involve multiple modalities, such as texts, images and videos… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
35
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 70 publications
(38 citation statements)
references
References 25 publications
0
35
1
Order By: Relevance
“…Indeed the community is aware of this form of abuse and there have been several attempts for multimodal analysis (Singh et al, 2017;Yang et al, 2019;Gomez et al, 2020). In our work, however, we do not address the aspect of multimodal abuse simply because many datasets only include the textual component of a micropost and the reconstruction of non-textual components of posts can only be reconstructed with greater effort or even not be obtained at all.…”
Section: Multimodal Abusementioning
confidence: 99%
“…Indeed the community is aware of this form of abuse and there have been several attempts for multimodal analysis (Singh et al, 2017;Yang et al, 2019;Gomez et al, 2020). In our work, however, we do not address the aspect of multimodal abuse simply because many datasets only include the textual component of a micropost and the reconstruction of non-textual components of posts can only be reconstructed with greater effort or even not be obtained at all.…”
Section: Multimodal Abusementioning
confidence: 99%
“…A large portion of the hateful content that is shared on social media is in the form of memes, which feature multiple modalities like audio, text, images and videos in some cases as well. [54] present different fusion approaches to tackle multi-modal information for hate speech detection. [11] explore multi-modal hate speech consisting of text and image modalities.…”
Section: Related Workmentioning
confidence: 99%
“…Most recent works have focused on leveraging neural networks in this task (Chen et al, 2015;Nguyen and Grishman, 2015;Nguyen et al, 2016;Ghaeini et al, 2016;Feng et al, 2016). The existing approaches can be categorized into two classes: The first class is to improve ED through special learning techniques including adversarial training (Hong et al, 2018), knowledge distillation (Liu et al, 2019; and model pretraining (Yang et al, 2019). The second class is to improve ED by introducing extra resource, such as argument information , document information (Duan et al, 2017;Chen et al, 2018), multi-lingual information (Liu et al, 2018a(Liu et al, , 2019, knowledge base and syntactic information (Sha et al, 2018).…”
Section: Related Workmentioning
confidence: 99%