2013 IEEE Conference on Computer Vision and Pattern Recognition 2013
DOI: 10.1109/cvpr.2013.477
|View full text |Cite
|
Sign up to set email alerts
|

Improving an Object Detector and Extracting Regions Using Superpixels

Abstract: We propose an approach to improve the detection performance of a generic detector when it is applied to a particular video. The performance of offline-trained objects detectors are usually degraded in unconstrained video environments due to variant illuminations, backgrounds and camera viewpoints. Moreover, most object detectors are trained using Haar-like features or gradient features but ignore video specific features like consistent color patterns. In our approach, we apply a Superpixel-based Bag-of-Words (… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0
2

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 67 publications
(52 citation statements)
references
References 14 publications
0
50
0
2
Order By: Relevance
“…As can be observed in Section 4.3, the overwhelming majority of the state-of-the-art research for domain adaptation of object detectors in videos use self-training in one form or another [29,30,31,32,33,36,37,38,40,42,11,10]. In order to adapt a generic pedestrian detector to a specific scene, a typical system would run the generic detector on some frames in a video, then score each detection using some heuristics and afterwards, add the most confident positive and negative detections to the original dataset for retraining.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…As can be observed in Section 4.3, the overwhelming majority of the state-of-the-art research for domain adaptation of object detectors in videos use self-training in one form or another [29,30,31,32,33,36,37,38,40,42,11,10]. In order to adapt a generic pedestrian detector to a specific scene, a typical system would run the generic detector on some frames in a video, then score each detection using some heuristics and afterwards, add the most confident positive and negative detections to the original dataset for retraining.…”
Section: Discussionmentioning
confidence: 99%
“…at least two) source domains whereas most of the works reviewed in this paper assume that only one source domain is available. The approach of Shu et al [42] may work poorly for videos with small pedestrians.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Instead, our goal is to train the best object detectors for a specific scenario. More recently, [20,21,22] improve generic offline trained detectors using specific scenarios.…”
Section: Related Workmentioning
confidence: 99%
“…In a comprehensive study, Dollár et al (2011) show that the recall rates of 16 different pedestrian detectors decrease rapidly if the intersection-over-union score threshold is increased. A better alignment of the detection result to the real object boundaries is for instance achieved by finer segmentation, based on pixels (Dai and Hoiem, 2012), superpixels (Shu et al, 2013), interest points (Ommer et al, 2009), (Gall and Lempitsky, 2013), object parts (Felzenszwalb et al, 2010), (Benfold and Reid, 2011) or contour models (Leibe et al, 2005), (Gavrila and Munder, 2007). Such models have the advantage of being more robust against partial occlusions compared to a holistic model.…”
Section: Introductionmentioning
confidence: 99%