2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00203
|View full text |Cite
|
Sign up to set email alerts
|

DPOD: 6D Pose Object Detector and Refiner

Abstract: In this paper we present a novel deep learning method for 3D object detection and 6D pose estimation from RGB images. Our method, named DPOD (Dense Pose Object Detector), estimates dense multi-class 2D-3D correspondence maps between an input image and available 3D models. Given the correspondences, a 6DoF pose is computed via PnP and RANSAC. An additional RGB pose refinement of the initial pose estimates is performed using a custom deep learning-based refinement scheme. Our results and comparison to a vast num… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
415
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 453 publications
(434 citation statements)
references
References 34 publications
(95 reference statements)
1
415
0
Order By: Relevance
“…Unlike the previous categories of methods, i.e., classification-based and regressionbased, this category performs the classification and regression tasks within a single architecture. The methods can firstly do the classification, the outcomes of which are cured in a regression-based refinement step [105], [84], [78], [166] or vice versa [75], or can do the classification and regression in a single-shot process [87], [145], [101], [106], [100], [148], [103], [102], [30], [37], [162].…”
Section: B Regressionmentioning
confidence: 99%
See 1 more Smart Citation
“…Unlike the previous categories of methods, i.e., classification-based and regressionbased, this category performs the classification and regression tasks within a single architecture. The methods can firstly do the classification, the outcomes of which are cured in a regression-based refinement step [105], [84], [78], [166] or vice versa [75], or can do the classification and regression in a single-shot process [87], [145], [101], [106], [100], [148], [103], [102], [30], [37], [162].…”
Section: B Regressionmentioning
confidence: 99%
“…DeepContext [78] is trained on the partially synthetic training depth images which exhibit a variety of different local object appearances, and real data are used to fine tune the method. The 2D-driven 3D methods and the 3D BB detectors work at the level of categories, and the 6D methods [30], [37], [166], [162] work at instance-level.…”
Section: Classification and Regressionmentioning
confidence: 99%
“…Another feature not well covered by synthetic data is proper illumination. Recent methods [21,29,16,56] prerender a number of synthetic images featuring different light conditions. Here, we instead implement differentiable lighting based on the simple Phong model [35], which is fully operated by the network.…”
Section: Light Module (L)mentioning
confidence: 99%
“…When it comes to deep learning methods, training detectors on real data yields the best results. However, the fact that 3D models of the objects are available and training data can be synthesized by rendering them has been used in only a few studies, most notably using such detectors as SSD6D [20], AAE [29] and DPOD [34]. It is remarkable that all deep learning 6DoF object detectors trained either on real or synthetic data use a single neural network per object, in contrast to 2D object detectors, such as YOLO [24], SSD [21] or R-CNNs [14,13,25,15], which use one network for all object classes.…”
Section: Introductionmentioning
confidence: 99%
“…In the following sections, besides describing the steps of the dataset creation pipeline, we also present detection and 6D pose estimation results for all the newly-defined benchmarks of one of a recently introduced methods -Dense Pose Object Detector (DPOD) [34]. This method is trained on all the objects at once or on all the objects present in the test scene on strictly synthetic renderings of provided 3D models.…”
Section: Introductionmentioning
confidence: 99%