The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.00740
|View full text |Cite
|
Sign up to set email alerts
|

Towards Part-Based Understanding of RGB-D Scans

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 45 publications
0
3
0
Order By: Relevance
“…To recover object shapes, some methods Groueix et al 2018;Wang et al 2018] reconstruct meshes from a template, and others [Huang et al 2018b;Izadinia et al 2017] adopt shape retrieval approaches to search from a given CAD database. Recently, some approaches [Dahnert et al 2021;Nie et al 2020;Popov et al 2020;Yang and Zhang 2016;Zhang et al 2021b] enable 3D scene understanding by generating a room layout, camera pose, object bounding boxes, or even meshes from a single view, automatically completing and annotating scene meshes [Bokhovkin et al 2021] or predicting object alignments and layouts [Avetisyan et al 2020] from an RGB-D scan. Inspired by PanoContext [Zhang et al 2014] that panoramic images contain richer context information than the perspective ones, Zhang et al [Zhang et al 2021a] propose a better 3D scene understanding method with panoramic captures as input.…”
Section: Related Workmentioning
confidence: 99%
“…To recover object shapes, some methods Groueix et al 2018;Wang et al 2018] reconstruct meshes from a template, and others [Huang et al 2018b;Izadinia et al 2017] adopt shape retrieval approaches to search from a given CAD database. Recently, some approaches [Dahnert et al 2021;Nie et al 2020;Popov et al 2020;Yang and Zhang 2016;Zhang et al 2021b] enable 3D scene understanding by generating a room layout, camera pose, object bounding boxes, or even meshes from a single view, automatically completing and annotating scene meshes [Bokhovkin et al 2021] or predicting object alignments and layouts [Avetisyan et al 2020] from an RGB-D scan. Inspired by PanoContext [Zhang et al 2014] that panoramic images contain richer context information than the perspective ones, Zhang et al [Zhang et al 2021a] propose a better 3D scene understanding method with panoramic captures as input.…”
Section: Related Workmentioning
confidence: 99%
“…In comparison, our focus is on learning part-based semantic and instance segmentation of noisy and fragmented real-world 3D scans. Very recently, initial approaches to semantic 3D segmentation have been proposed (Bokhovkin et al, 2021;Uy et al, 2019) but for a significantly less extensive part hierarchy. More specifically, (Bokhovkin et al, 2021) targets predicting part hierarchy at object and coarse parts levels, discarding smaller parts altogether; in contrast, we are able to predict parts at finer levels in the hierarchy.…”
Section: Related Workmentioning
confidence: 99%
“…In the domain of object modeling, part-based approaches leverage computer vision techniques to track movements among object parts [22,23], exploit contextual relations from large datasets [24][25][26] or develop data-efficient learning methods [27,28]. These approaches aim to recognize and segment object parts, enhancing the understanding of complex object structures, but they do not yield a holistic representation of a scene that encompasses multiple objects.…”
Section: A Related Workmentioning
confidence: 99%