2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412514
|View full text |Cite
|
Sign up to set email alerts
|

Two-Level Attention-based Fusion Learning for RGB-D Face Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
2
2

Relationship

1
5

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 22 publications
0
10
0
Order By: Relevance
“…The obtained embeddings were finally fused to feed an SVM classifier for performing FR. Jiang et al [38] presented an 2013 Hand-crafted HOG RDF Feature-level IIIT-D [25] 2013 Hand-crafted ICP, DCS SRC N/A IIIT-D [34] 2014 Hand-crafted PCA, LBP, SIFT, LGBP kNN Score-level Kinect Face [27] 2014 Hand-crafted RISE+HOG RDF Feature-level IIIT-D [24] 2016 Hand-crafted ICP SDF N/A Lock3DFace [35] 2016 Hand-crafted Covariance matrix rep. SVM Score-level CurtinFaces [28] 2016 Deep learning Autoencoder Softmax Score-level Kinect Face [36] 2018 Deep learning Siamese CNN Softmax Feature-level Pandora [30] 2018 Deep learning 9 Layers CNN + Inception Softmax Feature-level VAP, IIIT-D, Lock3DFace [37] 2018 Deep learning Fine-tuned VGG-Face Softmax Feature-level LFFD [38] 2018 Deep learning Custom CNN Attribute-aware loss Feature-level Private dataset [39] 2018 Deep learning Inception-v2 Softmax Feature-level IIIT-D, Lock3DFace [40] 2019 Deep learning 14 layers CNN + Attention Softmax Feature-level Lock3DFace [32] 2020 Deep learning CNN + two-level attention Softmax Feature-level IIIT-D, CurtinFaces [41] 2020 Deep learning Custom CNN Assoc., Discrim., and Softmax Feature-level IIIT-D attribute-aware loss function for CNN-based FR which aims to regularize the distribution of the learned feature vectors with respect to some soft-biometric attributes such as gender, ethnicity, and age, thus boosting FR results. Cui et al [39] estimated the depth from RGB modality using a multi-task approach including face identification along with depth estimation.…”
Section: A Rgb-d Face Recognition Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…The obtained embeddings were finally fused to feed an SVM classifier for performing FR. Jiang et al [38] presented an 2013 Hand-crafted HOG RDF Feature-level IIIT-D [25] 2013 Hand-crafted ICP, DCS SRC N/A IIIT-D [34] 2014 Hand-crafted PCA, LBP, SIFT, LGBP kNN Score-level Kinect Face [27] 2014 Hand-crafted RISE+HOG RDF Feature-level IIIT-D [24] 2016 Hand-crafted ICP SDF N/A Lock3DFace [35] 2016 Hand-crafted Covariance matrix rep. SVM Score-level CurtinFaces [28] 2016 Deep learning Autoencoder Softmax Score-level Kinect Face [36] 2018 Deep learning Siamese CNN Softmax Feature-level Pandora [30] 2018 Deep learning 9 Layers CNN + Inception Softmax Feature-level VAP, IIIT-D, Lock3DFace [37] 2018 Deep learning Fine-tuned VGG-Face Softmax Feature-level LFFD [38] 2018 Deep learning Custom CNN Attribute-aware loss Feature-level Private dataset [39] 2018 Deep learning Inception-v2 Softmax Feature-level IIIT-D, Lock3DFace [40] 2019 Deep learning 14 layers CNN + Attention Softmax Feature-level Lock3DFace [32] 2020 Deep learning CNN + two-level attention Softmax Feature-level IIIT-D, CurtinFaces [41] 2020 Deep learning Custom CNN Assoc., Discrim., and Softmax Feature-level IIIT-D attribute-aware loss function for CNN-based FR which aims to regularize the distribution of the learned feature vectors with respect to some soft-biometric attributes such as gender, ethnicity, and age, thus boosting FR results. Cui et al [39] estimated the depth from RGB modality using a multi-task approach including face identification along with depth estimation.…”
Section: A Rgb-d Face Recognition Methodsmentioning
confidence: 99%
“…Mu et al [40] proposed adding an attention weight map to each feature map, computed from RGB and depth modalities, thus focusing on the most important pixels with respect to their locations during training. Uppal et al [32] used both spatial and channel information from depth and RGB images and fused the information using a two-step attention mechanism. The attention modules assign weights to features, choosing between features from depth and RGB and hence utilize the information from both data modalities effectively.…”
Section: B Attention Mechanismsmentioning
confidence: 99%
See 3 more Smart Citations