2020 International Joint Conference on Neural Networks (IJCNN) 2020
DOI: 10.1109/ijcnn48605.2020.9206905
|View full text |Cite
|
Sign up to set email alerts
|

I didn’t mean what I wrote! Exploring Multimodality for Sarcasm Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(11 citation statements)
references
References 12 publications
0
11
0
Order By: Relevance
“…Since final predictions are made using features of the highlighted areas, this operation gives an overall boost to the model performance. Moreover, we can verify from Figure 3(a) and 3(b) that our model is able to attend to text like "pack of almonds" and "stupid people" without the need to explicitly use noisy Optical Character Recognition (OCR), an approach used in previous works [10,15].…”
Section: Model Analysismentioning
confidence: 77%
See 2 more Smart Citations
“…Since final predictions are made using features of the highlighted areas, this operation gives an overall boost to the model performance. Moreover, we can verify from Figure 3(a) and 3(b) that our model is able to attend to text like "pack of almonds" and "stupid people" without the need to explicitly use noisy Optical Character Recognition (OCR), an approach used in previous works [10,15].…”
Section: Model Analysismentioning
confidence: 77%
“…The work in [1] extracts visual features and visual attributes from images using ResNet and builds a hierarchical fusion model to detect sarcasm. Along the same lines, the recurrent network model in [15] proposes the idea of a gating mechanism to leak information from one modality to the other and achieves superior performance on Twitter benchmark dataset for sarcasm detection. The authors of [19] use pre-trained BERT and ResNet models to encode text and image data and connect the two using a gate called a bridge.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…modal 12 textual and bi-modal textual+acoustic information. In text modality, SVM on T avg reports mediocre F1-scores of 22.5% and 35.6% for the sarcasm and humor classification, respectively.…”
Section: Comparative Analysismentioning
confidence: 99%
“…Thus, it offers an excellent opportunity to study the sarcasm or humor in a context. A few previous attempts [10,11,12] on sarcasm classification involved multi-modal information in a conversation to leverage the context and extract the incongruity between the surface and expressed semantics. Similarly, many studies [13,14] employed images and visual frames along with the text to detect humor.…”
Section: Introductionmentioning
confidence: 99%