Proceedings of the 15th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Application 2020
DOI: 10.5220/0008877700360046
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Scene Completion from a Single 360-Degree Image and Depth Map

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(16 citation statements)
references
References 0 publications
0
16
0
Order By: Relevance
“…Note that datasets providing 3D meshes or point clouds easily be voxelized as detailed in [118]. Additionally, Stanford 2D-3D-S [3] provides 360 • RGB-D images, of interest for completing entire rooms [29]. Due to real datasets small sizes, low scene variability, and annotation ambiguities, synthetic SUNCG [118] (aka SUNCG-D) was proposed, being a large scale dataset with pairs of depth images and complete synthetic scene meshes.…”
Section: Datasetsmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that datasets providing 3D meshes or point clouds easily be voxelized as detailed in [118]. Additionally, Stanford 2D-3D-S [3] provides 360 • RGB-D images, of interest for completing entire rooms [29]. Due to real datasets small sizes, low scene variability, and annotation ambiguities, synthetic SUNCG [118] (aka SUNCG-D) was proposed, being a large scale dataset with pairs of depth images and complete synthetic scene meshes.…”
Section: Datasetsmentioning
confidence: 99%
“…Voxel grid encodes scene geometry as 3D grid, which cells describe semantic occupancy of the space. Opposed to point clouds, grids conveniently define neighborhood with adjacent cells, and thus enable easy application of 3D CNNs, which facilitates to extend deep learning architectures designed for 2D data into 3D [14,17,19,22,24,28,29,39,49,68,108,118,155,158]. However, the representation suffers from constraining limitations and efficiency drawbacks since it represents both occupied and free regions of the scene, leading to high memory and computation needs.…”
Section: Scene Representationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing works all use geometrical inputs like depth [12,25,[39][40][41][42]45], occupancy grids [13,25,55,69] or point cloud [53,81]. Truncated Signed Distance Function (TSDF) were also proved informative [6,9,10,12,20,21,41,59,64,77,79]. Among others originalities, some SSC works use adversarial training to guide realism [10,64], exploit multi-task [6,38], or use lightweight networks [40,55].…”
Section: Related Workmentioning
confidence: 99%
“…Semantic segmentation tasks aim to assign a semantic class label to every pixel in the input image. Examples of applications in scene understanding include PixelNet [11], which performs semantic segmentation and edge detection; EdgeNet [12], which combines depth information with semantic scene completion, using RGB-D input data. For synthetic data generation, UnrealCV provides a pipeline that generates images from VEs providing semantic segmentations [13], allowing for easy generation of training data.…”
Section: Related Workmentioning
confidence: 99%