2020 International Conference on Electronics, Information, and Communication (ICEIC) 2020
DOI: 10.1109/iceic49074.2020.9051332
|View full text |Cite
|
Sign up to set email alerts
|

Facial Expression Recognition in Videos: An CNN-LSTM based Model for Video Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(13 citation statements)
references
References 12 publications
1
12
0
Order By: Relevance
“…After this replacement, the accuracy of the video classifier is 56.8%. This is in line with state-of-the-art results in the literature on emotion recognition from RAVDESS videos, namely 57.5% with Synchronous Graph Neural Networks (8 emotions) [50]; 61% with ConvNet-LSTM (8 emotions) [1]; 59% with an RNN (7 emotions) [9], and 82.4% with stacked autoencoders (6 emotions) [5].…”
Section: A Dataset and Model Architecturesupporting
confidence: 88%
See 2 more Smart Citations
“…After this replacement, the accuracy of the video classifier is 56.8%. This is in line with state-of-the-art results in the literature on emotion recognition from RAVDESS videos, namely 57.5% with Synchronous Graph Neural Networks (8 emotions) [50]; 61% with ConvNet-LSTM (8 emotions) [1]; 59% with an RNN (7 emotions) [9], and 82.4% with stacked autoencoders (6 emotions) [5].…”
Section: A Dataset and Model Architecturesupporting
confidence: 88%
“…V, has been studied by other authors in-the-clear, i.e. without regards for privacy protection, using a variety of deep learning architectures, with reported accuracies in the 57%-82% range, depending on the number of emotion classes included in the study (6 to 8) [5], [50], [9], [1]. The ConvNet model that we trained for our experimental results in Sec.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…RAVDESS is classbalanced except the neutral class, which was elicited 50% less time than the other emotion classes. We adapted two crossvalidation settings following the methods [42], [48], [27], [28], [13], [72], [44], [53], [12], [52]. The first setting considers the identities of the actors such that the training (validation) and the corresponding testing k-folds have no overlap in terms of actors (shown as actor-split= hereafter).…”
Section: A Datasets and Evaluation Metricsmentioning
confidence: 99%
“…The majority of works mainly concentrated on unimodal learning of emotions [11], [12], [13], i.e., processing a single modality. Although there exist breakthrough achievements by unimodal emotion recognition, due to the aforementioned multimodal nature of emotion expression, such models remain incapable in some circumstances.…”
Section: Introductionmentioning
confidence: 99%