2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00375
|View full text |Cite
|
Sign up to set email alerts
|

3D-RCNN: Instance-Level 3D Object Reconstruction via Render-and-Compare

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
296
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 322 publications
(297 citation statements)
references
References 38 publications
0
296
0
Order By: Relevance
“…Several artificial intelligence (AI) research groups, including ours, have shown how neural network "inference models" can be built from a feedforward or recurrent network architecture trained to infer the underlying scene structure, rather than to recognize objects or classify object categories as in conventional DCNNs. In contrast to early analysis-by-synthesis algorithms, inference is fast, following a single bottom-up passes from the image or a small number of bottomup-top-down cycles, without the need for extensive iterative processing [24][25][26][27][28][29] . These models have been developed in an engineering setting, and are just beginning to be tested in machine vision problems; their correspondence with human perception or neural mechanisms is unexplored.…”
Section: Introductionmentioning
confidence: 99%
“…Several artificial intelligence (AI) research groups, including ours, have shown how neural network "inference models" can be built from a feedforward or recurrent network architecture trained to infer the underlying scene structure, rather than to recognize objects or classify object categories as in conventional DCNNs. In contrast to early analysis-by-synthesis algorithms, inference is fast, following a single bottom-up passes from the image or a small number of bottomup-top-down cycles, without the need for extensive iterative processing [24][25][26][27][28][29] . These models have been developed in an engineering setting, and are just beginning to be tested in machine vision problems; their correspondence with human perception or neural mechanisms is unexplored.…”
Section: Introductionmentioning
confidence: 99%
“…Afterwards, estimated mesh information is used to recover the allocentric orientation [36] of the target object. Egocentric orientation can then be recovered and lifted to 6D by adopting different approaches from the literature [22,19].…”
Section: Methodsmentioning
confidence: 99%
“…[36] learned the best 3D surface keypoints for pose estimation, pursuing keypoints geometrically and semantically consistent from different viewpoint. Whereas target objects in other problems are in a categorical level [14,36,22,1], deformable and/or articulated, objects in 6D pose estimation are typically in an instance level and rigid. In this paper, we propose a deep architecture that recovers an object's 6D pose from the input of a single 2D RGB image via 3D reconstruction.…”
Section: Introductionmentioning
confidence: 99%
“…Different from [53,24] where render-and-compare losses are computed upon silhouettes and 2D depth maps, we compute dense render-and-compare losses L rac using IUV values between ground-truth IUV images and rendered ones (see Sec. 3.4).…”
Section: Dense Render-and-comparementioning
confidence: 99%