Abstract:Face frontalization can boost the performance of face recognition methods and has made significant progress with the development of Generative Adversarial Networks (GANs). However, many GAN-based face frontalization methods still perform relatively weak on face recognition tasks under large face poses. In this paper, we propose Feature-Improving GAN (FI-GAN) for face frontalization, which aims to improve the recognition performance under large face poses. We assume that there is an inherent mapping between the… Show more
“…Frontal-Frontal Frontal-Profile Sengupta et al [3] 96.40 ± 0.69 84.91 ± 1.82 Sankarana et al [47] 96.93 ± 0.61 89.17 ± 2.35 LightCNN [42] 99.37 ± 0.30 91.56 ± 1.89 ResNet50 [43] 99.54 ± 0.31 94.25 ± 1.33 DR-GAN [12] 97.84 ± 0.79 93.41 ± 1.17 DR-GAN AM [19] 98.36 ± 0.75 93.89 ± 1.39 Chen et al [48] 98.67 ± 0.36 91.97 ± 1.70 PIM [17] 99.44 ± 0.36 93.10 ± 1.01 Peng et al [8] 98.67 ± − 93.76 ± − FI-GAN [32] 98…”
Section: Methodsmentioning
confidence: 99%
“…Their model first encode images by utilizing a pre-trained face expert network and then tried to recover photorealistic images from the extracted feature. Recently, Rong et al [32] proposed FI-GAN, aiming at improving the recognition performance under large face poses via a Feature-Mapping Block which maps the features of profile space to the frontal space. In most face frontalization methods, a large number of paired nonfrontal-frontal face images, typically from the MultiPIE dataset, are required to train the model in a fully supervised manner.…”
Section: Indoor Faces Forward Streamsmentioning
confidence: 99%
“…By contrast, most existing face frontalization methods are only trained on indoor face images (typically from Multi-PIE [20] dataset), which limits their generalization abilities in the unconstrained environment. [31] 93.62 ± 1.17 98.38 ± 0.06 HPEN [30] 96.25 ± 0.76 99.39 ± 0.02 LightCNN [42] 98.87 ± 0.61 99.69 ± 0.17 ResNet50 [43] 98.98 ± 0.52 99.79 ± 0.14 FF-GAN [13] 96.42 ± 0.89 99.45 ± 0.03 A3FCNN [15] 96.63 ± 0.99 99.29 ± 0.42 FI-GAN [32] 98. 30 We also conducted the face recognition experiment on the Multi-PIE dataset following the settings used with TP-GAN [14].…”
“…Frontal-Frontal Frontal-Profile Sengupta et al [3] 96.40 ± 0.69 84.91 ± 1.82 Sankarana et al [47] 96.93 ± 0.61 89.17 ± 2.35 LightCNN [42] 99.37 ± 0.30 91.56 ± 1.89 ResNet50 [43] 99.54 ± 0.31 94.25 ± 1.33 DR-GAN [12] 97.84 ± 0.79 93.41 ± 1.17 DR-GAN AM [19] 98.36 ± 0.75 93.89 ± 1.39 Chen et al [48] 98.67 ± 0.36 91.97 ± 1.70 PIM [17] 99.44 ± 0.36 93.10 ± 1.01 Peng et al [8] 98.67 ± − 93.76 ± − FI-GAN [32] 98…”
Section: Methodsmentioning
confidence: 99%
“…Their model first encode images by utilizing a pre-trained face expert network and then tried to recover photorealistic images from the extracted feature. Recently, Rong et al [32] proposed FI-GAN, aiming at improving the recognition performance under large face poses via a Feature-Mapping Block which maps the features of profile space to the frontal space. In most face frontalization methods, a large number of paired nonfrontal-frontal face images, typically from the MultiPIE dataset, are required to train the model in a fully supervised manner.…”
Section: Indoor Faces Forward Streamsmentioning
confidence: 99%
“…By contrast, most existing face frontalization methods are only trained on indoor face images (typically from Multi-PIE [20] dataset), which limits their generalization abilities in the unconstrained environment. [31] 93.62 ± 1.17 98.38 ± 0.06 HPEN [30] 96.25 ± 0.76 99.39 ± 0.02 LightCNN [42] 98.87 ± 0.61 99.69 ± 0.17 ResNet50 [43] 98.98 ± 0.52 99.79 ± 0.14 FF-GAN [13] 96.42 ± 0.89 99.45 ± 0.03 A3FCNN [15] 96.63 ± 0.99 99.29 ± 0.42 FI-GAN [32] 98. 30 We also conducted the face recognition experiment on the Multi-PIE dataset following the settings used with TP-GAN [14].…”
“…Some of the best performing DNN-based frontalization methods use CNN/GAN architectures, e.g. [54,27,49,59,57,43,58,55], which outperform CNNonly ones, e.g. [53].…”
Face frontalization consists of synthesizing a frontallyviewed face from an arbitrarily-viewed one. The main contribution of this paper is a robust frontalization method that preserves non-rigid facial deformations, i.e. expressions, to perform lip reading. The method iteratively estimates the rigid transformation (scale, rotation, and translation) and the non-rigid deformation between 3D landmarks extracted from an arbitrarily-viewed face, and 3D vertices parameterized by a deformable shape model. An important merit of the method is its ability to deal with large Gaussian and non-Gaussian errors in the data. For that purpose, we use the generalized Student-t distribution. The associated EM algorithm assigns a weight to each observed landmark, the higher the weight the more important the landmark, thus favoring landmarks that are only affected by rigid head movements. We propose to use the zero-mean normalized cross-correlation (ZNCC) score to evaluate the ability to preserve facial expressions. We show that the method, when incorporated into a deep lipreading pipeline, considerably improves the word classification score on an in-the-wild benchmark. An extended version with supplemental materials can be found at https: //team.inria.fr/robotlearn/rff-vsr/ * This work has been partially supported by MIAI@Grenoble Alpes (ANR-19-P3IA-0003).
“…W Ith the improvement of generative models, such as the Generative Adversarial Networks (GANs) [1]- [3] and variational autoencoders (VAEs) [4], the processing of image-to-image translation have made considerable progresses. The translation includes photo style translation [5], [6], objects dyeing [7]- [9], and also facial synthesis [10]- [13].…”
Recently, Generative Adversarial Network (GAN) based approaches are applied in facial attribute translation. However, many tasks, i.e. multi facial attributes translation and background invariance, are not well handled in the literature. In this paper, we propose a novel GAN-based method that aims to get the target image that performs better within modifying one or more facial attributes in a single model. The model generator learns multi-points by inputting a re-coded transfer vector, ensuring the single model could learn multiple attributes simultaneously. It also optimizes the cycle loss to enhance the efficiency of transferring multi attributes. Moreover, the method uses the adaptive parameter to improve the calculation method of the loss function of the residual image. The results are also compared with the StarGAN v2, which is the current state-of-the-art model to prove the effectiveness and advancedness. Experiments show that our method has a satisfactory performance in multi facial attributes translation.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.