2012
DOI: 10.1007/978-3-642-32717-9_3
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchy of Localized Random Forests for Video Annotation

Abstract: Abstract. We address the problem of annotating a video sequence with partial supervision. Given the pixel-wise annotations in the first frame, we aim to propagate these labels ideally throughout the whole video. While some labels can be propagated using optical flow, disocclusion and unreliable flow in some areas require additional cues. To this end, we propose to train localized classifiers on the annotated frame. In contrast to a global classifier, localized classifiers allow to distinguish colors that appea… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 22 publications
(22 reference statements)
0
6
0
Order By: Relevance
“…[111] proposes a joint propagation strategy with synthesized training samples. To better account for errors in label propagation, [64] proposes a hierarchy of local classifiers for this task and [5] leverages a mixture-of-tree model for temporal association. The work of [15] leverages label propagation as a data augmentation scheme and demonstrate improved performance on semantic segmentation.…”
Section: Methodsmentioning
confidence: 99%
“…[111] proposes a joint propagation strategy with synthesized training samples. To better account for errors in label propagation, [64] proposes a hierarchy of local classifiers for this task and [5] leverages a mixture-of-tree model for temporal association. The work of [15] leverages label propagation as a data augmentation scheme and demonstrate improved performance on semantic segmentation.…”
Section: Methodsmentioning
confidence: 99%
“…In (14), we show that the weight of each template is determined by the product of their reliability and discriminative strength.…”
Section: )Udph 1xpehumentioning
confidence: 99%
“…While some algorithms combine both color and shape cues [4], [5] for video segmentation, existing methods do not take the reliability of shape 1: judge the motion scale, and generate the target and template regions 2: determine the template grid scale via (12) 3: solve σ by substituting (13) into (14), then obtain p u obj (x) and p u bck (x) with the computed σ 4: compute the reliability of the template region in the u-th template Cu(x) via (16) 5: obtain the color probability Pc(x|F ) and total discriminative strength D(x) for pixel x via (16) and (17) Output: color probability Pc(x|F ) and discriminative strength D(x) model into consideration, which makes the integration results less accurate. In this work, we exploit the reliability of both color and shape models effectively to generate the shape and color prior for segmentation.…”
Section: Integrating Shape and Color Modelsmentioning
confidence: 99%
“…These methods either assume that we have annotations on a few frames [5,26] or leave the ambiguous areas to human labeling [9]. Galasso [13] applies an off-line trained model to label all the region trajectories belonging to pedestrians.…”
Section: Related Workmentioning
confidence: 99%
“…This technique has a broad range of applications such as surveillance [4], video compression G. Zhang ( ) · Z. Yuan · Y. Liu · L. Ma · N. Zheng Institute of Artificial Intelligence and Robotics, #28 West Xianning Road, Xi'an, Shaanxi, People's Republic of China e-mail: zhangtetsu@gmail.com [8], video retrieval [25], and action recognition [15]. According to the applications, video segmentation can be described as followings: removing background from a video clip [33], tracking and extracting specific objects [23], propagating sparse annotations [5,26], and over-segmenting a video into spatiotemporal consistent super-voxels [38,39]. However, compared with image segmentation, video segmentation has more challenges and it is far less sophisticated for industrial needs.…”
Section: Introductionmentioning
confidence: 98%