Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.440
|View full text |Cite
|
Sign up to set email alerts
|

MMGCN: Multimodal Fusion via Deep Graph Convolution Network for Emotion Recognition in Conversation

Abstract: Emotion recognition in conversation (ERC) is a crucial component in affective dialogue systems, which helps the system understand users' emotions and generate empathetic responses. However, most works focus on modeling speaker and contextual information primarily on the textual modality or simply leveraging multimodal information through feature concatenation. In order to explore a more effective way of utilizing both multimodal and long-distance contextual information, we propose a new model based on multimod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
28
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 108 publications
(32 citation statements)
references
References 24 publications
1
28
0
Order By: Relevance
“…DialogueGCN (Ghosal et al, 2019) captures conversational dependencies between utterances with a graph-based structure. MMGCN (Hu et al, 2021) further proposes a GCNbased multimodal fusion method for multimodal ERC tasks to improve recognition performance. Di-alogXL (Shen et al, 2020) first introduces a strong pre-trained language model XLNet for text-based ERC.…”
Section: Related Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…DialogueGCN (Ghosal et al, 2019) captures conversational dependencies between utterances with a graph-based structure. MMGCN (Hu et al, 2021) further proposes a GCNbased multimodal fusion method for multimodal ERC tasks to improve recognition performance. Di-alogXL (Shen et al, 2020) first introduces a strong pre-trained language model XLNet for text-based ERC.…”
Section: Related Methodsmentioning
confidence: 99%
“…MMGCN: A state-of-the-art GCN-based multimodal ERC framework proposed in (Hu et al, 2021). For the uni-modal experiments, we only model the fully connected graph.…”
Section: Baseline Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…Moreover, DAG-ERC also makes meaningful and reasonable assumptions while constructing the graph by 1) Removing the link of an utterance in a dialogue to future utterances and 2) By imputing remote information for modeling conversational context by introducing another edge to the speakers previous utterance. Very recently MMGCN [15] proposed fusing information from multiple modalities by the use of spectral domain GCN to encode the multimodal contextual information. The work closest to our work is [11], where the authors use discourse relations between utterances to build a conversational graph and show that ER in both multi-party and two-party conversations benefit from conversational discourse structures.…”
Section: Graph-based Modelsmentioning
confidence: 99%