2021
DOI: 10.1016/j.specom.2020.12.001
|View full text |Cite
|
Sign up to set email alerts
|

Fusion of deep learning features with mixture of brain emotional learning for audio-visual emotion recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
10

Relationship

0
10

Authors

Journals

citations
Cited by 36 publications
(19 citation statements)
references
References 25 publications
0
19
0
Order By: Relevance
“…For multimodal fusion, one of the major challenges is how to effectively integrate data from different sources and design moderate architecture to complete representation learning. According to the fusion level or location, fusion concept can be divided into three types: data-level, featurelevel and decision-level [24]. As the development of the machine learning, feature-level and decision-level fusion (or called early and late fusion) have spawned a variety of studies.…”
Section: Related Workmentioning
confidence: 99%
“…For multimodal fusion, one of the major challenges is how to effectively integrate data from different sources and design moderate architecture to complete representation learning. According to the fusion level or location, fusion concept can be divided into three types: data-level, featurelevel and decision-level [24]. As the development of the machine learning, feature-level and decision-level fusion (or called early and late fusion) have spawned a variety of studies.…”
Section: Related Workmentioning
confidence: 99%
“…The OFC evaluates the response of the amygdala based on the input received from the Sensory cortex, which leads to the prevention of improper learning connections. In a nutshell, the Sensory cortex integrates the features extracted from different unimodal inputs and the amygdala and OFC works to form a decision after interacting with the memory [17,18]. BEL models require rewards extracted from input data and are derived from monotonic reinforcement learning.…”
Section: Brain Emotion Learning Modelsmentioning
confidence: 99%
“…When signals propagate through the network, a conditional layer learns which expert networks to activate, and so that the various combinations of experts are flexible under different circumstances. It was proven to outperform popular fusion strategies in dynamic emotion prediction using visual-audio [49], [50]. For developing DMoE model, we apply fully-connected layers for both expert and gating networks.…”
Section: Multimodal Fusion Modulesmentioning
confidence: 99%