2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298776
|View full text |Cite
|
Sign up to set email alerts
|

Unconstrained realtime facial performance capture

Abstract: We introduce a realtime facial tracking system specifically designed for performance capture in unconstrained settings using a consumer-level RGB-D sensor. Our framework provides uninterrupted 3D facial tracking, even in the presence of extreme occlusions such as those caused by hair, hand-to-face gestures, and wearable accessories. Anyone's face can be instantly tracked and the users can be switched without an extra calibration step. During tracking, we explicitly segment face regions from any occluding parts… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
68
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
4
1

Relationship

2
7

Authors

Journals

citations
Cited by 102 publications
(68 citation statements)
references
References 41 publications
0
68
0
Order By: Relevance
“…Generic blendshape models are used by some face tracking methods from monocular RGB video [Garrido et al 2013] or RGB-D video [Weise et al 2011] but need to be deformed into a static face scan or a set of scanned static expressions of an actor prior to tracking. Such generic blendshape adaptation fails to capture person-specific expression details, which is why some recent approaches estimate identity and blendshape parameters from captured face animations, and also person-specific correctives on top of this generic face model [Bouaziz et al 2013;Li et al 2013;Hsieh et al 2015]. However, all these approaches require RGB-D camera input.…”
Section: Related Workmentioning
confidence: 94%
“…Generic blendshape models are used by some face tracking methods from monocular RGB video [Garrido et al 2013] or RGB-D video [Weise et al 2011] but need to be deformed into a static face scan or a set of scanned static expressions of an actor prior to tracking. Such generic blendshape adaptation fails to capture person-specific expression details, which is why some recent approaches estimate identity and blendshape parameters from captured face animations, and also person-specific correctives on top of this generic face model [Bouaziz et al 2013;Li et al 2013;Hsieh et al 2015]. However, all these approaches require RGB-D camera input.…”
Section: Related Workmentioning
confidence: 94%
“…They use the depth and color information of an RGB-D camera to robustly segment the face region. Image taken from [HMYL15]. Figure 8: An RGB face tracker that is robust to occlusion has been proposed by [SLL16].…”
Section: Handling Occlusionsmentioning
confidence: 99%
“…In general, they are based on a segmentation mask that disables the data fitting terms in occluded regions. Hsieh et al [HMYL15] use depth and color information of an RGB-D camera to robustly segment the visible face region (see Fig. 7).…”
Section: Handling Occlusionsmentioning
confidence: 99%
“…An adaptive scheme was proposed to capture more detail with point-to-point deformation on top of blendshapes in [22]. To explicitly deal with outliers caused by occlusions, a method was proposed to segment the face and complete the occluded parts based on the blendshape in [19], which was later extended to RGB input in [28]. Binocular stereo system, on the other hand, can provide higher resolution and work in outdoor environments directly under sunlight, but are more prone to suffer from lighting variation.…”
Section: Expression Clusteringmentioning
confidence: 99%