2019
DOI: 10.1016/j.patrec.2017.10.022
|View full text |Cite
|
Sign up to set email alerts
|

Deep spatial-temporal feature fusion for facial expression recognition in static images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
52
1

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 90 publications
(56 citation statements)
references
References 12 publications
0
52
1
Order By: Relevance
“…However, for the situations with unavailability of a neutral face, static techniques for FER have an edge over dynamic techniques. To circumvent this, "average human face" was suggested as an effective alternative representation of neutral face in a dynamic model [24]. Hitherto, there is no consensus about one technique being superior to the other comprehensively.…”
Section: Methods For Analysis Of the Described Features: Dynamic And mentioning
confidence: 99%
“…However, for the situations with unavailability of a neutral face, static techniques for FER have an edge over dynamic techniques. To circumvent this, "average human face" was suggested as an effective alternative representation of neutral face in a dynamic model [24]. Hitherto, there is no consensus about one technique being superior to the other comprehensively.…”
Section: Methods For Analysis Of the Described Features: Dynamic And mentioning
confidence: 99%
“…They found that emotion classification was very susceptible to head orientation and illumination variation. Sun et al [31] proposed a multi-channel deep neural network that learned and fused the spatial-temporal features. Lopes et al [32] proposed several pre-processing steps before feeding the faces to a convolutional neural network for facial expression recognition.…”
Section: Visual Signal Based Emotion Classificationmentioning
confidence: 99%
“…These metrics driven by multiple features are subsequently learned with these extracted multiple features in a unified fashion to use complementary and discriminative data for emotion classification. Sun et al [20] introduced a multi-channel deep neural network that learns and puts together the spatialtemporal descriptors for facial expressions identification in static frames. The important concept of the algorithm is to discover and collect optical flow from the difference among the peak expression face frame and the neutral face frame as the temporal data of a specific facial expression, and consider the grey-level frame of peak expression face as the spatial data.…”
Section: Literature Reviewmentioning
confidence: 99%