2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00590
|View full text |Cite
|
Sign up to set email alerts
|

Attention-Driven Cropping for Very High Resolution Facial Landmark Detection

Abstract: Facial landmark detection is a fundamental task for many consumer and high-end applications and is almost entirely solved by machine learning methods today. Existing datasets used to train such algorithms are primarily made up of only low resolution images, and current algorithms are limited to inputs of comparable quality and resolution as the training dataset. On the other hand, high resolution imagery is becoming increasingly more common as consumer cameras improve in quality every year. Therefore, there is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
38
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 64 publications
(39 citation statements)
references
References 48 publications
0
38
0
Order By: Relevance
“…Implementation Details In our experiments, we first use the four stacked hourglass networks as our backbone [37]. All the training and testing images are In the next section, we firstly compare our algorithm with the state-of-theart methods, such as 3DDFA [44], RAR [29], Wing [36], LAB [8], DU-Net [48], Liu et al [11], ODN [9], HRNet [49], Chandran et al [13] and LUVLi [12].…”
Section: Dataset and Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…Implementation Details In our experiments, we first use the four stacked hourglass networks as our backbone [37]. All the training and testing images are In the next section, we firstly compare our algorithm with the state-of-theart methods, such as 3DDFA [44], RAR [29], Wing [36], LAB [8], DU-Net [48], Liu et al [11], ODN [9], HRNet [49], Chandran et al [13] and LUVLi [12].…”
Section: Dataset and Implementation Detailsmentioning
confidence: 99%
“…For instance, the coordinate regression facial landmark detection methods [6,8,9] learn features from the whole face images and then regress to the landmark coordinates, which drives the models to learn the whole facial features in a common/normal way that cannot accurately model the differences of local details and the relationships among local details. Also, the heatmap regression facial landmark detection methods [10,11,12,13] generate a landmark heatmap for each landmark and then predict landmarks by traversing the corresponding landmark heatmaps.…”
Section: Introductionmentioning
confidence: 99%
“…With the fast development of deep learning techniques in computer vision, deep learning based methods [9,11,21,27,46,47] have significantly boosted and outperformed both the template fitting method and cascaded regression-based method, creating a new state-of-the-art in facial landmark detection task. Most of them leverage deep convolutional neural networks (CNN) to learn facial features and predict the facial landmark in an end-to-end fashion.…”
Section: Mouth Landmark Detectionmentioning
confidence: 99%
“…Recently, landmark detection on faces and bodys has attracted great attention from researchers, and methods in this area can be roughly divided into five branches: detection from global input to local regions [7,8], direct localization [9], auxiliary inputs [10,11,12], and feature refinement [13,14,15,16].…”
Section: Landmark Detectionmentioning
confidence: 99%
“…Specifically, landmarks are detected with downsampled images using an hourglass network and refined with cropped local regions from input images using similar models [7]; low-resolution response maps are generated using a fully convolutional network, and high-resolution ones over local regions are produced to output more accurate landmarks [8]. CenterNet directly outputs landmarks from one center keypoint heatmap and two corner ones with a convolutional backbone [9].…”
Section: Landmark Detectionmentioning
confidence: 99%