ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2022
DOI: 10.1109/icassp43922.2022.9746418
|View full text |Cite
|
Sign up to set email alerts
|

Mmlatch: Bottom-Up Top-Down Fusion For Multimodal Sentiment Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 23 publications
(5 citation statements)
references
References 36 publications
0
5
0
Order By: Relevance
“…To separately address the relationships between each pair of modalities, BBFN integrates two bimodal fusion modules along with a gated control mechanism. Given that top-down interactions remain unaccounted in previous approaches, MMLatch (Paraskevopoulos, Georgiou, and Potamianos 2022) addresses this limitation by incorporating a feedback mechanism in the forward pass.…”
Section: Related Work Multimodal Fusion Methodsmentioning
confidence: 99%
“…To separately address the relationships between each pair of modalities, BBFN integrates two bimodal fusion modules along with a gated control mechanism. Given that top-down interactions remain unaccounted in previous approaches, MMLatch (Paraskevopoulos, Georgiou, and Potamianos 2022) addresses this limitation by incorporating a feedback mechanism in the forward pass.…”
Section: Related Work Multimodal Fusion Methodsmentioning
confidence: 99%
“…Kumar et al [39] achieved deep multimodal feature vector fusion by introducing learnable gating mechanisms, self-attended context representations, and recurrent layer-based self and gated cross-fusion. Paraskevopoulos et al [17] proposed a neural architecture for multimodal fusion, utilizing a feedback mechanism in the forward pass during network training to capture top-down cross-modal interactions. Subsequently, numerous studies have employed even more novel approaches.…”
Section: Multimodal Sentiment Analysismentioning
confidence: 99%
“…MMLATCH [17]: Bottom-up, top-down fusion proposes a neural architecture that captures top-down cross-modal interactions by using a feedback mechanism in the forward pass during network training.…”
Section: Baselinesmentioning
confidence: 99%
See 1 more Smart Citation
“…Hazarika et al designed a new framework that projects modalities into modality-invariant and modality-specific subspaces to achieve a more holistic view of the multimodal data [36]. Paraskevopoulos et al [37] introduced a neural architecture that adeptly captures cross-modal interactions from a top-down perspective to analyze users' sentiment. Transformer-based methods have also been proposed for MSA tasks, such as the multi-layer fusion module based on the transformer-encoder developed by Li et al [38], which incorporates contrastive learning to further explore sentiment features, and the text-enhanced transformer fusion model proposed by Wang et al to better understand text-oriented pairwise cross-modal mappings and acquire crucial unified multimodal representations [39].…”
Section: Multimodal Sentiment Analysismentioning
confidence: 99%