CVPR 2011 2011
DOI: 10.1109/cvpr.2011.5995448
|View full text |Cite
|
Sign up to set email alerts
|

From 3D scene geometry to human workspace

Abstract: We present a human-centric paradigm for scene understanding. Our approach goes beyond estimating 3D scene geometry and predicts the "workspace" of a human which is represented by a data-driven vocabulary of human interactions. Our method builds upon the recent work in indoor scene understanding and the availability of motion capture data to create a joint space of human poses and scene geometry by modeling the physical interactions between the two. This joint space can then be used to predict potential human p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
176
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 214 publications
(180 citation statements)
references
References 17 publications
2
176
0
Order By: Relevance
“…Prior works have explored learning based approaches for estimating the depth of the scene [33,36,40,43,65,79,81,94,98] and estimating depth ordering [41,45] from a single image for 3D scene understanding. Recent work has shown that we can also use the objects (clutter) in the scene to aid better depth estimation of the scene [34,37] using affordances.…”
Section: Static Camera Scenariomentioning
confidence: 99%
“…Prior works have explored learning based approaches for estimating the depth of the scene [33,36,40,43,65,79,81,94,98] and estimating depth ordering [41,45] from a single image for 3D scene understanding. Recent work has shown that we can also use the objects (clutter) in the scene to aid better depth estimation of the scene [34,37] using affordances.…”
Section: Static Camera Scenariomentioning
confidence: 99%
“…Scene Understanding: Geometry, Humans and Objects. Physical aspects of affordances have been recently explored in [10,12,5,16]. For example, interactions between a sitting human pose and the environment are used to identify sittable regions [10], semantic and geometric affordances of large objects such as furniture are extracted by observing people [5,12], and spatial affordances of objects with respect to possible human poses are used for placing and labeling objects [17,16].…”
Section: Related Workmentioning
confidence: 99%
“…recently, physical aspects have been explored in [10,12,16], where they model the functionality-based spatial interactions of humans with their environments. For example, Grabner et al [10] uses the interactions between a sitting human pose and the environment to identify sittable regions, Delaitre et al [5] observed people to extract semantic and geometric affordances for large furniture-like objects, and Jiang, Koppula and Saxena [16] uses the spatial affordances of objects with respect to possible human poses for the task of labeling objects.…”
Section: Introductionmentioning
confidence: 99%
“…The ability to infer the geometry of a scene has enabled a variety of applications in both the vision and graphics fields. For example, Gupta et al [8] use coarse geometry estimates to predict what locations in an environment afford various actions. Karsch et al [16] use scene geometry to realistically render additional objects into a scene.…”
Section: Introductionmentioning
confidence: 99%
“…However, because this approach aims to match the exact configuration of objects in an image, with an identical furniture configuration from a library of 3D models, the algorithm does not have the flexibility required to precisely reconstruct the diversity of object configurations found in natural scenes. Additionally, many scene understanding approaches such as [7,8,28] make limiting assumptions regarding the robustness of existing monocular autocalibration algorithms. When this preliminary stage of their pipelines fail, the algorithms are unable to recover and produce a reasonable result.…”
Section: Introductionmentioning
confidence: 99%