2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00314
|View full text |Cite
|
Sign up to set email alerts
|

Pix3D: Dataset and Methods for Single-Image 3D Shape Modeling

Abstract: Figure 1: We present Pix3D, a new large-scale dataset of diverse image-shape pairs. Each 3D shape in Pix3D is associated with a rich and diverse set of images, each with an accurate 3D pose annotation to ensure precise 2D-3D alignment. In comparison, existing datasets have limitations: 3D models may not match the objects in images; pose annotations may be imprecise; or the dataset may be relatively small. AbstractWe study 3D shape modeling from a single image and make contributions to it in three aspects. Firs… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
381
0
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 369 publications
(382 citation statements)
references
References 61 publications
(92 reference statements)
0
381
0
1
Order By: Relevance
“…Reconstructing real-world objects. To qualitatively evaluate the generalization performance of our method on the real images, we test our network on the Pix3D [28] dataset by using the model trained on the ShapeNet [3]. Figure 6 shows the results reconstructed by our method and AtlasNet, where the objects in the images are manually segmented.…”
Section: Comparisons With the State-of-the-artsmentioning
confidence: 99%
“…Reconstructing real-world objects. To qualitatively evaluate the generalization performance of our method on the real images, we test our network on the Pix3D [28] dataset by using the model trained on the ShapeNet [3]. Figure 6 shows the results reconstructed by our method and AtlasNet, where the objects in the images are manually segmented.…”
Section: Comparisons With the State-of-the-artsmentioning
confidence: 99%
“…To make it easier for PCDNet to serve as a baseline in subsequent research, we reported two common metric scores which are CD and intersection over union (IoU). CD is our main criterion, not because PCDNet is trained using CD, but it is better correlated with human perception [27]. IoU quantifies the overlapping region between two input sets.…”
Section: Resultsmentioning
confidence: 99%
“…Tatarchenko et al [79], Lin et al [69], and Sun et al [10] also estimate the binary/silhouette masks, along with the depth maps. The binary masks have been used to filter out points that are not backprojected to the surface in 3D space.…”
Section: Intermediatingmentioning
confidence: 99%