2021
DOI: 10.3389/fcomp.2021.636094
|View full text |Cite
|
Sign up to set email alerts
|

Systematic Evaluation of Design Choices for Deep Facial Action Coding Across Pose

Abstract: The performance of automated facial expression coding is improving steadily. Advances in deep learning techniques have been key to this success. While the advantage of modern deep learning techniques is clear, the contribution of critical design choices remains largely unknown, especially for facial action unit occurrence and intensity across pose. Using the The Facial Expression Recognition and Analysis 2017 (FERA 2017) database, which provides a common protocol to evaluate robustness to pose variation, we sy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…When developing an algorithm, the design is informed by the robustness to angle changes. Niimura et al [55] evaluated robustness to pose variation, including pitch, of algorithms for "deep facial action coding" according to pre-training, feature alignment, model size, and optimizers. Many computer vision researchers have attempted to develop pose normalization techniques [56,57].…”
Section: Discussionmentioning
confidence: 99%
“…When developing an algorithm, the design is informed by the robustness to angle changes. Niimura et al [55] evaluated robustness to pose variation, including pitch, of algorithms for "deep facial action coding" according to pre-training, feature alignment, model size, and optimizers. Many computer vision researchers have attempted to develop pose normalization techniques [56,57].…”
Section: Discussionmentioning
confidence: 99%
“…Then we synthesized 5,000 images for these pairs. We selected 5,000 images according to experimental results by Niinuma et al (2021b). They analyzed the influence of training set size on FERA17, and showed that the training set size have a minor influence on the performance: score peaked at 5,000 images, and after that performance plateaued.…”
Section: Train Au Classifier Without Personalizationmentioning
confidence: 99%
“…We selected a VGG16 network pre-trained on ImageNet for the baseline architecture for AU estimation. Previous studies found this combination preferable for AU coding (Niinuma et al, 2021b). We replaced the final layer of the network with a 6-length one-hot representation, and fine-tuned VGG16 network from the third convolutional layer.…”
Section: Train Au Classifier Without Personalizationmentioning
confidence: 99%
See 1 more Smart Citation