2022
DOI: 10.3390/app12031093
|View full text |Cite
|
Sign up to set email alerts
|

FMFN: Fine-Grained Multimodal Fusion Networks for Fake News Detection

Abstract: As one of the most popular social media platforms, microblogs are ideal places for news propagation. In microblogs, tweets with both text and images are more likely to attract attention than text-only tweets. This advantage is exploited by fake news producers to publish fake news, which has a devasting impact on individuals and society. Thus, multimodal fake news detection has attracted the attention of many researchers. For news with text and image, multimodal fake news detection utilizes both text and image … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 33 publications
(14 citation statements)
references
References 27 publications
0
14
0
Order By: Relevance
“…Wang et al [25] implemented a fine-grained multimodal fusion network (FMFN) as a nuanced method for the fusion of textual and visual features for effective FND. The suggested approach used scaled dot-product attention method to enhance both features and fuse the improved features, thereby capturing the province among features.…”
Section: Literature Surveymentioning
confidence: 99%
“…Wang et al [25] implemented a fine-grained multimodal fusion network (FMFN) as a nuanced method for the fusion of textual and visual features for effective FND. The suggested approach used scaled dot-product attention method to enhance both features and fuse the improved features, thereby capturing the province among features.…”
Section: Literature Surveymentioning
confidence: 99%
“…Year Method COOLANT [16] 2023 Cross-modal contrastive learning MMFN [4] 2023 Multi-grained information fusion MPFN [17] 2023 Multimodal progressive fusion SAMPLE [18] 2023 Similarity-aware multimodal prompt learning TieFake [19] 2023 Integration of multimodal context and author sentiment: focusing on title-text similarity and emotion awareness FNR [20] 2023 Similarity and transformer-based learning DGM [21] 2023 Transformer based on manipulation-aware contrastive learning and modality-aware cross-attention CAFE [22] 2022 Cross-modal ambiguity learning CMC [23] 2022 Cross-modal knowledge distillation FND-CLIP [24] 2022 Contrastive language-image pretraining-guided learning LIIMR [25] 2022 Leveraging intra and inter modality relationship FMFN [26] 2022 Fine-grained multimodal fusion network MCAN [27] 2021 Multimodal co-attention networks AMFB [28] 2021 Attention-based multimodal factorized bilinear pooling HMCAN [29] 2021 Hierarchical multi-modal contextual attention network CARMN [30] 2021 Crossmodal attention residual and multichannel CNN SAFE [31] 2020 Cross-modal similarity measurement stages or layers of a neural network. The main idea is to enable the model to progressively integrate information from different modalities, instead of attempting to do so in a single step.…”
Section: Modelmentioning
confidence: 99%
“…AMFB [28] proposes an attention-based stacked bi-directional LSTM network that captures textual information at different levels, attentionbased multi-level convolutional neural network-recurrent neural network for visual feature extraction, multimodal factorized bilinear pooling to combine the textual and visual feature representations, and then passes them through a multi-layer perceptron for FND. Wang et al [26] introduce a Fine-grained Multimodal Fusion Network (FMFN) to fuse textual and visual features fully. The proposed model utilizes scaled dotproduct attention mechanisms for the fine-grained fusion of the textual and the visual features, which not only takes into account the correlations between different visual features but also captures the dependencies between textual and visual features.…”
Section: Modelmentioning
confidence: 99%
“…TRANSFAKE adopts a preprocessing method similar to BERT for concatenated text, comment and image. In another work (Wang, Mao, and Li 2022), Wang et al apply scaled dot-product attention on top of image and text features as a fine-grained fusion and use the fused feature to classify articles. Wang et al propose a deep learning network for Biomedical informatics that leverages visual and textual information and a semantic-and tasklevel attention mechanism to focus on the essential contents of a post that signal anti-vaccine messages (Wang, Yin, and Argyris 2021).…”
Section: Concatenation-based Architecturesmentioning
confidence: 99%