2018
DOI: 10.1007/s11263-018-1077-3
|View full text |Cite
|
Sign up to set email alerts
|

Synthesizing a Scene-Specific Pedestrian Detector and Pose Estimator for Static Video Surveillance

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(14 citation statements)
references
References 63 publications
0
14
0
Order By: Relevance
“…The PET 2006 dataset includes multi-view camera sequences containing left-luggage scenarios at a train station in which the scene complexity increases. To evaluate the pedestrian detection performance, we used only a single viewpoint so that the evaluation would be performed under the same conditions as that for the comparison algorithms [36].…”
Section: Expeimental Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…The PET 2006 dataset includes multi-view camera sequences containing left-luggage scenarios at a train station in which the scene complexity increases. To evaluate the pedestrian detection performance, we used only a single viewpoint so that the evaluation would be performed under the same conditions as that for the comparison algorithms [36].…”
Section: Expeimental Resultsmentioning
confidence: 99%
“…To verify the effectiveness of the soft target training scheme, we compared its performance with that of six state-of-theart methods: (1) DPM [8]; (2) the Faster R-CNN approach, which shares full-image convolutional features with the detection network [12]; (3) the scene pose CNN network (SPN), which generates a scene-specific pedestrian detector and pose estimator [36]; (4) YOLO 9000, which is a realtime CNN-based object detection system over 9000 object categories [14]; (5) teacher RFs consisting of 300 trees (teacher RF); and (6) proposed S-RF consisting of 50 trees (proposed S-RF). Faster R-CNN and YOLO 9000 used pretrained model parameters without performing fine-tuning.…”
Section: B Detection Comparison On Pets2006 Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…From there, the use of synthetic visual data generated from virtual environments has kept growing. We found works using synthetic data for object detection/recognition [66][67][68][69], object viewpoint recognition [70], re-identification [71], and human pose estimation [72]; building synthetic cities for autonomous driving tasks such as semantic segmentation [44,73], place recognition [74], object tracking [45,75], object detection [76,77], stixel computation [78], and benchmarking different on-board computer vision tasks [47]; building indoor scenes for semantic segmentation [79], as well as normal and depth estimation [80]; generating GT for optical flow, scene flow, and disparity [81,82]; generating augmented reality images to support object detection [83]; simulating adverse atmospheric conditions such as rain or fog [84,85]; even performing procedural generation of videos for human action recognition [86,87]. Moreover, since robotics and autonomous driving rely on sensorimotor models worthy of being trained and tested dynamically, in the last years, the use of simulators has been intensified beyond datasets [48,49,88,89].…”
Section: Related Workmentioning
confidence: 99%
“…Dhome et al used synthetic models to recognize objects from a single image [11]. For pedestrian detection, computer generated pedestrian images were used to train classifiers [5]. 3D simulation has been used for multi-view car detection [1] [31] [6].…”
Section: Related Workmentioning
confidence: 99%