2021
DOI: 10.1109/lsp.2021.3078698
|View full text |Cite
|
Sign up to set email alerts
|

Interactive Multimodal Attention Network for Emotion Recognition in Conversation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(9 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…To verify the effectiveness and validity of the proposed hierarchical interactive IMC system, a series of experiments are carried out by comparing the proposed system with the state-of-the-art methods [34][35][36][37][38][39][40][41][42][43][44][45][46][47] for video sentiment analysis. Notably, the references are selected according to the following four criteria, i.e., content relevance, total cited times, journal academic impact, and timeliness.…”
Section: Application In Video Sentiment Analysismentioning
confidence: 99%
“…To verify the effectiveness and validity of the proposed hierarchical interactive IMC system, a series of experiments are carried out by comparing the proposed system with the state-of-the-art methods [34][35][36][37][38][39][40][41][42][43][44][45][46][47] for video sentiment analysis. Notably, the references are selected according to the following four criteria, i.e., content relevance, total cited times, journal academic impact, and timeliness.…”
Section: Application In Video Sentiment Analysismentioning
confidence: 99%
“…Tsai et al [35] used transformers to extract and fuse the features of the three modalities, which efectively solved the problems of unalignment of modal data and long-term dependencies between diferent modalities. Also, recently, pretrained networks using transfer learning techniques have achieved good performance for extracting features [36], especially in the feld of emotion recognition [37][38][39][40], and have advanced signifcantly. As the pretrained model can learn about global features from data, its parameters show better generalization efects.…”
Section: Related Workmentioning
confidence: 99%
“…Feature Fusion. At first, SCCA is used to fuse the features of the two modalities from facial expression and speech [20][21][22]. e SCCA algorithm can be expressed as follows:…”
Section: Classroom Psychological Assessmentmentioning
confidence: 99%
“…First, the proposed method extracts the features of facial expression and speech, respectively. en, the sparse canonical correlation analysis (SCCA) [20][21][22] algorithm is used to fuse the two kinds of features to obtain a unified feature. Finally, the sparse representation-based classification (SRC) is used for bimodal emotion recognition.…”
Section: Introductionmentioning
confidence: 99%