2021 IEEE Winter Conference on Applications of Computer Vision (WACV) 2021
DOI: 10.1109/wacv48630.2021.00245
|View full text |Cite
|
Sign up to set email alerts
|

Facial Expression Recognition in the Wild via Deep Attentive Center Loss

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
81
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 197 publications
(81 citation statements)
references
References 26 publications
0
81
0
Order By: Relevance
“…Wang et al [36] proposed a Region Attention Network (RAN) to capture facial regions for occlusion and pose variant FER. Farzaneh and Qi [11] introduced a Deep Attentive Center Loss (DACL) method to estimate the attention weight for the features for enhancing the discrimination. A sparse center loss was designed to achieve intra-class compactness and inter-class separation with the weighted features.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Wang et al [36] proposed a Region Attention Network (RAN) to capture facial regions for occlusion and pose variant FER. Farzaneh and Qi [11] introduced a Deep Attentive Center Loss (DACL) method to estimate the attention weight for the features for enhancing the discrimination. A sparse center loss was designed to achieve intra-class compactness and inter-class separation with the weighted features.…”
Section: Related Workmentioning
confidence: 99%
“…We set the batch size to 100 with a learning rate of 4 × 10 −5 . Unlike many methods [11,19] that rely on complicated loss, we use the standard label smoothing cross-entropy loss. Evaluation on RAF-DB: Table 1 compares POSTER with previous methods on the RAF-DB dataset.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…At the testing time, the central crops are fed into a trained model on AffectNet-7 and DAF_DB for 60 epochs; the initial learning rate is 0.01 with a reduction in a factor of 10 every 20 epochs; the batch size is set to 32 for AffectNet-7 and 16 for DAF_DB, and in experiments, the model is trained with Pytorch on NVIDIA GeForce GTX 1650 GPU with 16 GB RAM. Notably, the pre-training strategy is adapted for saving the total training time and can obtain superior performance [ 51 ]. In this paper, the proposed model will be pre-trained on a face dataset MS-CELEB-1M [ 52 ], and then fine-tuned on FER datasets, owing to the similarity between the domain of FER and the face recognition (FR) task.…”
Section: Resultsmentioning
confidence: 99%
“…Accuracy, % 8 classes 7 classes PSR (VGG-16) [20] 60.68 -ARM (ResNet-18) [21] 59.75 64.49 RAN [22] 59.5 -PAENet [8] -65.29 DACL [23] -65.20 EmotionNet (InceptionResNet-v1) [8] -64.74 CPG [24] - dataset. Moreover, we report the results of models after the first training stage, in which only the weights of the classification head were learned, while the other part of the model remains the same as in the pre-trained face recognition CNN (left part of Fig.…”
Section: Methodsmentioning
confidence: 99%