2018
DOI: 10.1109/tpami.2018.2791608
|View full text |Cite
|
Sign up to set email alerts
|

EAC-Net: Deep Nets with Enhancing and Cropping for Facial Action Unit Detection

Abstract: In this paper, we propose a deep learning based approach for facial action unit (AU) detection by enhancing and cropping regions of interest of face images. The approach is implemented by adding two novel nets (a.k.a. layers): the enhancing layers and the cropping layers, to a pretrained convolutional neural network (CNN) model. For the enhancing layers (noted as E-Net), we have designed an attention map based on facial landmark features and apply it to a pretrained neural network to conduct enhanced learning.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
68
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 101 publications
(68 citation statements)
references
References 10 publications
0
68
0
Order By: Relevance
“…Jaiswal and Valstar (2016) use a pre-defined binary mask created to select a relevant region for a particular AU and pass it to a convolutional and bidirectional Long Short-Term Memory (LSTM) neural network. Li et al (2018) design an attention map using the facial key points and AU centers to enforce their CNN-based architecture to focus more on these AU centers. Sanchez et al (2018) generate heatmaps for a target AU, by estimating the facial landmarks and drawing a 2D Gaussian around the points where the AU is known to cause changes.…”
Section: Regional Attentionmentioning
confidence: 99%
See 2 more Smart Citations
“…Jaiswal and Valstar (2016) use a pre-defined binary mask created to select a relevant region for a particular AU and pass it to a convolutional and bidirectional Long Short-Term Memory (LSTM) neural network. Li et al (2018) design an attention map using the facial key points and AU centers to enforce their CNN-based architecture to focus more on these AU centers. Sanchez et al (2018) generate heatmaps for a target AU, by estimating the facial landmarks and drawing a 2D Gaussian around the points where the AU is known to cause changes.…”
Section: Regional Attentionmentioning
confidence: 99%
“…Adversarial Training Framework (ATF) (Zhang et al, 2018) is a CNN-based framework in which AU loss is minimized and identity loss is maximized to learn subject invariant feature representations during the adversarial training. Finetuned VGG Network (FVGG) (Li et al, 2018) is the model obtained after finetuning the pretrained VGG 19-layer model. Network with enhancing layers (E-Net) (Li et al, 2018) is the finetuned VGG network with enhancing layer which forces the network to pay more attention to AU interest regions on face images.…”
Section: Performance Comparison With the State-of-the-artmentioning
confidence: 99%
See 1 more Smart Citation
“…The contributions focused on a diverse set of topics, techniques, and applications. Estimation of facial markers (e.g., facial landmarks, action units) was the most approached task in papers within the special issue [4], [5], [6], [7]. In fact, most papers in the issue used facial landmarks in one way or another as part of their methodology.…”
Section: The Computational Face Sectionmentioning
confidence: 99%
“…A similar methodology was adopted by Wang et al who combined a super resolution CNN and a recurrent neural net to iteratively refine landmarks in a cascade regression approach [6]. A novel deep architecture was proposed by Li et al in which attention maps and cropping layers were included in a CNN for better detection of facial action units [7]. In contrast to previous work, Wang et al described an iterative methodology based on deformable parts models for person-specific landmark detection.…”
Section: The Computational Face Sectionmentioning
confidence: 99%