2019
DOI: 10.15546/aeei-2018-0029
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation of Depth Modality in Convolutional Neural Network Classification of RGB-D Images

Abstract: This paper investigates the value of depth modality in object classification in RGB-D images. We use a simple model based on a multi-layered convolutional neural network which we train on a dataset of segmented RGB-D images of household and office objects. We evaluate and quantify the benefit of additional depth modality and its effect on classification accuracy on this dataset. Also, we compare the benefit of depth channel against the addition of color to grayscale image. Our experimental results support a co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 11 publications
(21 reference statements)
0
2
0
Order By: Relevance
“…Inspired by our findings of the significance of the spatial (depth) information during real-life object classification and scene understanding [37], we decided to focus on the classification of 3D images. Generally, these classification methods can be grouped into view-based and volume, or shape-based techniques, depending on the form of data that are present at the input of the classifier.…”
Section: Related Workmentioning
confidence: 99%
“…Inspired by our findings of the significance of the spatial (depth) information during real-life object classification and scene understanding [37], we decided to focus on the classification of 3D images. Generally, these classification methods can be grouped into view-based and volume, or shape-based techniques, depending on the form of data that are present at the input of the classifier.…”
Section: Related Workmentioning
confidence: 99%
“…To be usable as an input for 3D convolution-based networks it has to be converted to a voxel grid format. For this purpose our data transformation pipeline has been used [18].…”
Section: Preprocessing and Transformationmentioning
confidence: 99%