2021
DOI: 10.3390/app11167217
|View full text |Cite
|
Sign up to set email alerts
|

Guided Spatial Transformers for Facial Expression Recognition

Abstract: Spatial Transformer Networks are considered a powerful algorithm to learn the main areas of an image, but still, they could be more efficient by receiving images with embedded expert knowledge. This paper aims to improve the performance of conventional Spatial Transformers when applied to Facial Expression Recognition. Based on the Spatial Transformers’ capacity of spatial manipulation within networks, we propose different extensions to these models where effective attentional regions are captured employing fa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 35 publications
(44 reference statements)
0
5
0
Order By: Relevance
“…For the facial emotion recognizer, we adjusted the weights of the pre-trained STN on the AffectNet dataset to apply transfer-learning strategies. The trained STN for sentiment recognition on AffectNet reached an accuracy of 70.60%, as we can see in [60]. However, applying Feature-Extraction and max.…”
Section: Facial Emotion Recognition Resultsmentioning
confidence: 58%
See 3 more Smart Citations
“…For the facial emotion recognizer, we adjusted the weights of the pre-trained STN on the AffectNet dataset to apply transfer-learning strategies. The trained STN for sentiment recognition on AffectNet reached an accuracy of 70.60%, as we can see in [60]. However, applying Feature-Extraction and max.…”
Section: Facial Emotion Recognition Resultsmentioning
confidence: 58%
“…Therefore, we trained the STN again with the same database using seven emotions, the same as RAVDESS except for the 'Calm' emotion. This second model reached an accuracy of 65.90% on the AffectNet database using the same parameters and evaluation strategy as in [60].…”
Section: Facial Emotion Recognition Resultsmentioning
confidence: 94%
See 2 more Smart Citations
“…Where ( , ) is the target coordinates of the normal grid of the output feature map, ( , ) is the source coordinates of the input feature map that defines the sample points, and is the affine transformation matrix which might take various transformations. Normalized height and width coordinates so that #1 $ % & , % & $ 1 are within the spatial boundaries of the output and #1 $ % ( , % ( $ 1 are within the spatial boundaries of the input [20]. We can take this mechanism and combine it with ResNet, as seen in Fig.…”
Section: B Proposed Methodsmentioning
confidence: 99%