2019
DOI: 10.3390/app9122478
|View full text |Cite
|
Sign up to set email alerts
|

Model-Based 3D Pose Estimation of a Single RGB Image Using a Deep Viewpoint Classification Neural Network

Abstract: This paper presents a model-based approach for 3D pose estimation of a single RGB image to keep the 3D scene model up-to-date using a low-cost camera. A prelearned image model of the target scene is first reconstructed using a training RGB-D video. Next, the model is analyzed using the proposed multiple principal analysis to label the viewpoint class of each training RGB image and construct a training dataset for training a deep learning viewpoint classification neural network (DVCNN). For all training images … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 47 publications
(78 reference statements)
0
3
0
Order By: Relevance
“…In most studies, a sparse 3D model consisting of an image set of scenes and 3D points (sparse 3D model) has been used for deep learning-based camera pose estimation [12][13][14][15]19]. Specifically, 2D-3D matching has been investigated in coarse-to-fine localization studies focusing on improving the performance of deep learning methods [14,15].…”
Section: Proposed Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…In most studies, a sparse 3D model consisting of an image set of scenes and 3D points (sparse 3D model) has been used for deep learning-based camera pose estimation [12][13][14][15]19]. Specifically, 2D-3D matching has been investigated in coarse-to-fine localization studies focusing on improving the performance of deep learning methods [14,15].…”
Section: Proposed Methodsmentioning
confidence: 99%
“…The target of these studies is mainly the robotic manipulation of specific objects, such as chairs and boxes, and human-robot interaction. Deep neural networks have also been used for 6-DoF tracking of rigid objects in large occluded environments [16,19]. However, because these methods require red, green, blue, and depth (RGB-D) sensing data, their practical use is constrained to indoor scenes.…”
Section: Introductionmentioning
confidence: 99%
“…Some methods, as described for pose estimation, also use RGB-D cameras to perform data fusion for tracking [58][59][60]. We can use a system designed for single- [61] or multi-object [62] tracking that can perform the tracking of unknown objects (without any a priori information) [63,64] or use the object information (e.g., size or color) [65,66] to develop our tracking algorithms. Currently, most applied pose tracking methods rely on deep learning methods to obtain better performances without having to hand-craft features that can discriminate the tracked object on the search space (image frame when using a RGB camera).…”
Section: Pose Trackingmentioning
confidence: 99%