2015 IEEE International Conference on Robotics and Automation (ICRA) 2015
DOI: 10.1109/icra.2015.7139363
|View full text |Cite
|
Sign up to set email alerts
|

RGB-D object recognition and pose estimation based on pre-trained convolutional neural network features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
181
0
2

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 275 publications
(193 citation statements)
references
References 10 publications
2
181
0
2
Order By: Relevance
“…Asif et al [1] report improved recognition performance using a cascade of Random Forest classifiers that are fused in a hierarchical manner. Finally, in recent independent work Schwarz et al [20] proposed to use features extracted from CNNs pre-trained on ImageNet for RGB-D object recognition. While they also make use of a two-stream network they do not fine-tune the CNN for RGB-D recognition, but rather just use the pre-trained network as is.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Asif et al [1] report improved recognition performance using a cascade of Random Forest classifiers that are fused in a hierarchical manner. Finally, in recent independent work Schwarz et al [20] proposed to use features extracted from CNNs pre-trained on ImageNet for RGB-D object recognition. While they also make use of a two-stream network they do not fine-tune the CNN for RGB-D recognition, but rather just use the pre-trained network as is.…”
Section: Related Workmentioning
confidence: 99%
“…[14] 77.7 ± 1.9 78.8 ± 2.7 86.2 ± 2.1 CKM Desc. [3] N/A N/A 86.4 ± 2.3 CNN-RNN [22] 80.8 ± 4.2 78.9 ± 3.8 86.8 ± 3.3 Upgraded HMP [5] 82.4 ± 3.1 81.2 ± 2.3 87.5 ± 2.9 CaRFs [1] N/A N/A 88.1 ± 2.4 CNN Features [20] 83.1 ± 2.0 N/A 89.4 ± 1.3 Ours, Fus-CNN (HHA) 84.1 ± 2.7 83.0 ± 2.7 91.0 ± 1.9 Ours, Fus-CNN (jet) 84.1 ± 2.7 83.8 ± 2.7 91.3 ± 1.4 a preliminary experiment. A fixed momentum value of 0.9 and a mini-batch size of 128 was used for all experiments if not stated otherwise.…”
Section: A Experimental Setupmentioning
confidence: 99%
“…Features from the colour and depth channels were learned separately and then concatenated for use in the final softmax classifier. Schwarz et al 25 proposed using two pretrained CNNs to extract features from colour and depth images individually. Then, Eitel et al 12 proposed a similar structure to that in the method of Schwarz et al 25 The difference was that, in the latter, the fusion CNNs were trained end-to-end using the RGB-D data, which gives a higher accuracy.…”
Section: Related Workmentioning
confidence: 99%
“…Schwarz et al 25 proposed using two pretrained CNNs to extract features from colour and depth images individually. Then, Eitel et al 12 proposed a similar structure to that in the method of Schwarz et al 25 The difference was that, in the latter, the fusion CNNs were trained end-to-end using the RGB-D data, which gives a higher accuracy. Bai et al 26 proposed dividing the input images into several subsets according to their shapes and colours.…”
Section: Related Workmentioning
confidence: 99%
“…While most methods use 2D visual information only [2], there are numerous 3D shape based recognition techniques [3,4], as well as methods that use both visual and shape information [5,6]. Object detection methods are essential for scene understanding [7], which has a number of applications in different fields, such as robotics [1] or augmented reality [8].…”
Section: Introductionmentioning
confidence: 99%