2017 IEEE International Conference on Computer Vision Workshops (ICCVW) 2017
DOI: 10.1109/iccvw.2017.109
|View full text |Cite
|
Sign up to set email alerts
|

RGB-D Object Recognition Using Deep Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 39 publications
(24 citation statements)
references
References 17 publications
0
24
0
Order By: Relevance
“…The problem of estimating the depth map of a scene from its single image has attracted a lot of attention in the field of computer vision, since depth maps have a lot of applications, such as augmented reality [22], human computer interaction [33], human activity recognition [12], scene recognition [41], and segmentation [27,32]. The recent employment of convolutional neural networks (CNNs) has accelerated the research of the problem [9,25,23,2,4].…”
Section: Introductionmentioning
confidence: 99%
“…The problem of estimating the depth map of a scene from its single image has attracted a lot of attention in the field of computer vision, since depth maps have a lot of applications, such as augmented reality [22], human computer interaction [33], human activity recognition [12], scene recognition [41], and segmentation [27,32]. The recent employment of convolutional neural networks (CNNs) has accelerated the research of the problem [9,25,23,2,4].…”
Section: Introductionmentioning
confidence: 99%
“…Compared deep models demonstrated relative improvements in range of 3% to 10% (see Table 1). The evaluation has been performed using our CNN-based classifier along with results published in [7], [11] and [4] acquired using several different architectures. The comparison shows that the benefit of depth modality is comparable across multiple different classification methods.…”
Section: Resultsmentioning
confidence: 99%
“…Some models, like VGG3D [11] try to utilize depth information by generating RGB voxel grids from RGB-D images. These are then fed into a modified VGGnet model which is pre-trained on a large 2D RGB dataset.…”
Section: Classification Of 3d Datamentioning
confidence: 99%
“…As an example, given all trained models, the combination of different outputs may also further improve the denoising performance. It should also be noted that Zia et al proposed to extend a 2D filter to a 3D one by replicating this 2D filter along the depth dimension for RGB-D object recognition [45]. However, that extension cannot guarantee that the initial 3D network has the same performance as the trained 2D network, i.e.…”
Section: Discussionmentioning
confidence: 99%