2020 International Joint Conference on Neural Networks (IJCNN) 2020
DOI: 10.1109/ijcnn48605.2020.9206776
|View full text |Cite
|
Sign up to set email alerts
|

Attention-based 3D Object Reconstruction from a Single Image

Abstract: Recently, learning-based approaches for 3D reconstruction from 2D images have gained popularity due to its modern applications, e.g., 3D printers, autonomous robots, self-driving cars, virtual reality, and augmented reality. The computer vision community has applied a great effort in developing functions to reconstruct the full 3D geometry of objects and scenes. However, to extract image features, they rely on convolutional neural networks, which are ineffective in capturing long-range dependencies. In this pa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 18 publications
0
7
0
Order By: Relevance
“…Reconstruction based on single-view images: Realistically, multi-view images from abundant object instances are difficult to acquire. For that reason, recent work like Salvi [26] introduced learning-based architectures for 3D shape reconstruction using solely single input images. Thanks to the differentiable renderer [2,18,21], several frameworks [2,7,15,18,19,21] were presented to bridge the gap between an input image and its resulting texture by utilizing differentiable rendering and image reconstruction.…”
Section: Related Workmentioning
confidence: 99%
“…Reconstruction based on single-view images: Realistically, multi-view images from abundant object instances are difficult to acquire. For that reason, recent work like Salvi [26] introduced learning-based architectures for 3D shape reconstruction using solely single input images. Thanks to the differentiable renderer [2,18,21], several frameworks [2,7,15,18,19,21] were presented to bridge the gap between an input image and its resulting texture by utilizing differentiable rendering and image reconstruction.…”
Section: Related Workmentioning
confidence: 99%
“…316 well. However, as [8] showed in some of their experiments, the benefit from self-attention 396 modules was highest when used at the early layers of the encoder as opposed to the later 397 layers.…”
mentioning
confidence: 92%
“…-Bednarik et al [3] and Patch-Net [4] for reconstruction of depth and normal maps using a real dataset of texture-less surfaces, -HDM-Net [5] and IsMo-GAN [6], which reconstruct 3D point clouds from a synthetic dataset of textured surfaces, and -Pixel2Mesh [7], Salvi et al [8] and Yuan et al [9] for reconstruction of mesh-based models using a subset of the ShapeNet [10].…”
mentioning
confidence: 99%
See 2 more Smart Citations