2018
DOI: 10.1186/s13640-018-0262-1
|View full text |Cite
|
Sign up to set email alerts
|

Coarse-to-fine online learning for hand segmentation in egocentric video

Abstract: Hand segmentation is one of the most fundamental and crucial steps for egocentric human-computer interaction. The special egocentric view brings new challenges to hand segmentation tasks, such as the unpredictable environmental conditions. The performance of traditional hand segmentation methods depend on abundant manually labeled training data. However, these approaches do not appropriately capture the whole properties of egocentric human-computer interaction for neglecting the user-specific context. It is on… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 29 publications
0
5
0
Order By: Relevance
“…Usually, methods for online hand segmentation made assumptions on the hand motion [55], [56], [57], [58] and/or required the user to perform a calibration with pre-defined hand movements [59]. In this way, the combination of color and motion features facilitates the detection of hand pixels, in order to train segmentation models online.…”
Section: Lack Of Pixel-level Annotationsmentioning
confidence: 99%
See 3 more Smart Citations
“…Usually, methods for online hand segmentation made assumptions on the hand motion [55], [56], [57], [58] and/or required the user to perform a calibration with pre-defined hand movements [59]. In this way, the combination of color and motion features facilitates the detection of hand pixels, in order to train segmentation models online.…”
Section: Lack Of Pixel-level Annotationsmentioning
confidence: 99%
“…These matches were estimated using RANSAC [61] and after being removed, those left were assumed to belong to the hands and used to locate the seed point for region growing. Zhao et al [57], [58] based their approach on the typical motion pattern during actions involving the hands: preparatory phase (i.e., the hands move from the lower part of the frame to the image center) and interaction phase. During the preparatory phase they used a motion-based segmentation, computing the TV-L1 optical flow [62].…”
Section: Lack Of Pixel-level Annotationsmentioning
confidence: 99%
See 2 more Smart Citations
“…Beside body joints, keypoints can be extended to refer to the small visual units with semantic information indicating the compositions, shapes and poses of the target objects, such as finger joints or key positions of any other objects. Therefore, accurate keypoint detection in unconstrained environments brings benefit to other more detailed visual understanding tasks, including semantic segmentation[1, 2, 3], saliency object segmentation [4,5,6], hand segmentation [7] and pose estimation [8,9], viewpoint estimation [10,11,12,13], salient object detection [14,15,16], attention prediction [17] and 3D reconstruction [18,19,20]. Similar as many computer vision tasks, the progress on human pose estimation problem is significantly improved by deep convolutional neural networks.…”
Section: Introductionmentioning
confidence: 99%