2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01289
|View full text |Cite
|
Sign up to set email alerts
|

Putting People in their Place: Monocular Regression of 3D People in Depth

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
63
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
5

Relationship

4
6

Authors

Journals

citations
Cited by 102 publications
(63 citation statements)
references
References 40 publications
0
63
0
Order By: Relevance
“…In early times, Han et al [4] proposed to compress the silhouettes of the sequence in the temporal dimension, thus obtaining Gait Energy Image (GEI). Recently, due to the success of deep learning on multimedia and computer vision tasks [5,20,21,27,33,[36][37][38][39][40], deep learning-based methods also dominated the performance of gait recognition. For example, Shiraga et al [25] and Wu et al [32] proposed to learn gait features from GEIs by CNNs and significantly outperformed previous methods.…”
Section: Related Work 21 Gait Recognitionmentioning
confidence: 99%
“…In early times, Han et al [4] proposed to compress the silhouettes of the sequence in the temporal dimension, thus obtaining Gait Energy Image (GEI). Recently, due to the success of deep learning on multimedia and computer vision tasks [5,20,21,27,33,[36][37][38][39][40], deep learning-based methods also dominated the performance of gait recognition. For example, Shiraga et al [25] and Wu et al [32] proposed to learn gait features from GEIs by CNNs and significantly outperformed previous methods.…”
Section: Related Work 21 Gait Recognitionmentioning
confidence: 99%
“…Early methods are focused on retrieval-based pipelines to retarget the subject in a video to the desired motion with optical flow [11] or 3D human skeleton [41]. Recent methods usually adopt deep neural networks, especially GANs [13], and exploit 2D human pose [2,5,30,35,36], human parsing [8,21,26,39], 3D human pose [37], and other 3D information [20,22,31,44] for human motion transfer [23,24].…”
Section: Related Work 21 Human Motion Transfermentioning
confidence: 99%
“…Although monocular human pose and shape estimation [15,44] has been extensively explored over the past years, estimating global space locations together with human poses and shapes for multiple people from a single image is still a difficult problem due to the depth ambiguity. Existing methods [13,37] reconstruct 3D poses, shapes and relative positions of the reconstructed human meshes by assuming a constant focal length. But the methods are limited to small scenes with a common FoV (Field of View).…”
Section: Introductionmentioning
confidence: 99%