Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022) 2022
DOI: 10.18653/v1/2022.semeval-1.88
|View full text |Cite
|
Sign up to set email alerts
|

TeamOtter at SemEval-2022 Task 5: Detecting Misogynistic Content in Multimodal Memes

Abstract: We describe our system for the SemEval 2022 task on detecting misogynous content in memes. This is a pressing problem and we explore various methods ranging from traditional machine learning to deep learning models such as multimodal transformers. We propose a multimodal BERT architecture that uses information from both image and text. We further incorporate common world knowledge from pretrained CLIP and Urban dictionary. We also provide qualitative analysis to support out model. Our best performing model ach… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 10 publications
0
3
0
Order By: Relevance
“…Table 10 Continued from previous page Leaderboard Sub-task B Team Name Weighted-average F 1 -score 13 TeamOtter* (Maheshwari and Nangi, 2022) 0.680 14 Tathagata Raha* (Raha et al, 2022) 0.679 15 UPB* (Paraschiv et al, 2022) 0…”
Section: C2 Leader-board Of Sub-task Bmentioning
confidence: 99%
See 1 more Smart Citation
“…Table 10 Continued from previous page Leaderboard Sub-task B Team Name Weighted-average F 1 -score 13 TeamOtter* (Maheshwari and Nangi, 2022) 0.680 14 Tathagata Raha* (Raha et al, 2022) 0.679 15 UPB* (Paraschiv et al, 2022) 0…”
Section: C2 Leader-board Of Sub-task Bmentioning
confidence: 99%
“…Leaderboard Sub-task A (Srivastava, 2022) 0.759 R2D2* (Sharma et al, 2022b) 0.757 PAIC (ZHI et al, 2022) 0.755 ymf924 0.755 RubCSG* (Yu et al, 2022) 0.755 hate-alert 0.753 AMS_ADRN* (Li et al, 2022) 0.746 TIB-VA* (Hakimov et al, 2022) 0.734 4 union 0.727 Unibo* (Muti et al, 2022) 0.727 MMVAE* (Gu et al, 2022b) 0.723 YMAI* (Habash et al, 2022) 0.722 Transformers* (Mahadevan et al, 2022) 0.718 taochen* (Tao and jae Kim, 2022) 0.716 codec* (Mahran et al, 2022) 0.715 QMUL* 0.714 UPB* (Paraschiv et al, 2022) 0.714 HateU* (Arango et al, 2022) 0.712 yuanyuanya 0.708 Triplo7* (Attanasio et al, 2022) 0.699 InfUfrgs* (Lorentz and Moreira, 2022) 0.698 Mitra Behzadi* (Behzadi et al, 2022) 0.694 Gini_us* 0.692 5 riziko 0.687 UMUTeam* (García-Díaz et al, 2022) 0.687 Tathagata Raha* (Raha et al, 2022) 0.687 LastResort* (Agrawal and Mamidi, 2022) 0.686 TeamOtter* (Maheshwari and Nangi, 2022) 0.679 ShailyDesai 0.677 JRLV* (Ravagli and Vaiani, 2022) 0.670 I2C* (Cordon et al, 2022) 0.665 qinian* (Gu et al, 2022a) 0…”
mentioning
confidence: 99%
“…To assess the effectiveness of the textbased approach, the authors conducted training on a range of language models, which included RNN-based models, RoBERTa, and Ernie-2.0. A multimodal BERT architecture that utilises images and text data by incorporating world knowledge from pre-trained CLIP and Urban dictionary is discussed by Maheshwari et al [36]. For classification tasks based on visual and textual features, simple machine learning baselines, namely SVM, Naive Bayes, and Logistic Regression models are being used.…”
Section: Related Workmentioning
confidence: 99%