2020
DOI: 10.1109/access.2020.3023871
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Subject Multimodal Emotion Recognition Based on Hybrid Fusion

Abstract: Multimodal emotion recognition has gained traction in affective computing research community to overcome the limitations posed by the processing a single form of data and to increase recognition robustness. In this study, a novel emotion recognition system is introduced, which is based on multiple modalities including facial expressions, galvanic skin response (GSR) and electroencephalogram (EEG). This method follows a hybrid fusion strategy and yields a maximum one-subject-out accuracy of 81.2% and a mean acc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
42
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 98 publications
(56 citation statements)
references
References 88 publications
2
42
1
Order By: Relevance
“…Taking into account that the brain regions related to the frontal lobe have high recognition accuracy [28], the 6-channel EEG signals of the forehead and the PPS signals of other remaining channels are used as experimental data in the experiment. The data is downsampled to 128 Hz, and five bands including the delta (4-8 Hz), theta (8-13 Hz), alpha (13)(14)(15)(16)(17)(18)(19)(20)(21)(22)(23)(24)(25)(26)(27)(28)(29)(30), beta (30)(31)(32)(33)(34)(35)(36)(37)(38)(39)(40)(41)(42)(43), and gamma bands (4-43 Hz) are filtered out. Due to the error in the first 3 s of the video in the experiment, the first 3 s of the video are removed, and the middle 30 s of the remaining duration of the video are used as experimental data.…”
Section: A Data Set Settingsmentioning
confidence: 99%
See 1 more Smart Citation
“…Taking into account that the brain regions related to the frontal lobe have high recognition accuracy [28], the 6-channel EEG signals of the forehead and the PPS signals of other remaining channels are used as experimental data in the experiment. The data is downsampled to 128 Hz, and five bands including the delta (4-8 Hz), theta (8-13 Hz), alpha (13)(14)(15)(16)(17)(18)(19)(20)(21)(22)(23)(24)(25)(26)(27)(28)(29)(30), beta (30)(31)(32)(33)(34)(35)(36)(37)(38)(39)(40)(41)(42)(43), and gamma bands (4-43 Hz) are filtered out. Due to the error in the first 3 s of the video in the experiment, the first 3 s of the video are removed, and the middle 30 s of the remaining duration of the video are used as experimental data.…”
Section: A Data Set Settingsmentioning
confidence: 99%
“…Multimodal emotion recognition aims to combine the predictive capabilities of individual behavioral trails and biometric features for accurate classification [15]. The challenges of multimodal emotion recognition are as follows:…”
Section: Introductionmentioning
confidence: 99%
“…Since the early 2010s, deep neural networks, such as deep Boltzmann machine and autoencoder, have been extensively applied in order for multimodal data to effectively learn a fusion of features [42], [73]. Particularly, in domains where the data normally exist in multiple modalities, such as an activity, context recognition [74], pose estimation [75], emotion recognition [76], [77], and medical diagnosis [43], [78], deep neural networks have proven superior to other methods. Various kinds of deep neural networks are employed because they can handle multiple modalities effortlessly.…”
Section: Multimodal Learning In a Manufacturing Domainmentioning
confidence: 99%
“…Early data fusion, which integrates multiple sources of data into one feature vector, cannot fully exploit the complementary relationships among the data. Decision fusion based on the modality independence assumption makes the final decision based on multiple classifiers, but this assumption is often not true in practice (Cimtay et al , 2020). Feature fusion, which aims to combine each modality feature into one feature set, has been widely used in various classification tasks.…”
Section: Literature Reviewmentioning
confidence: 99%