2019
DOI: 10.1007/978-3-030-28603-3_8
|View full text |Cite
|
Sign up to set email alerts
|

RGB-D Image-Based Object Detection: From Traditional Methods to Deep Learning Techniques

Abstract: Object detection from RGB images is a long-standing problem in image processing and computer vision. It has applications in various domains including robotics, surveillance, human-computer interaction, and medical diagnosis. With the availability of low cost 3D scanners, a large number of RGB-D object detection approaches have been proposed in the past years. This chapter provides a comprehensive survey of the recent developments in this field. We structure the chapter into two parts; the focus of the first pa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 85 publications
(194 reference statements)
0
5
0
Order By: Relevance
“…Models such as ShapeNets [7] directly work with 3D voxel data. Data collected using depth sensors can be presented as RGB-D data [8][9][10] or point clouds [11,12] representing the topology of features present. Often, depth information is sparsely collected due to limitations of the depth sensors themselves.…”
Section: Related Workmentioning
confidence: 99%
“…Models such as ShapeNets [7] directly work with 3D voxel data. Data collected using depth sensors can be presented as RGB-D data [8][9][10] or point clouds [11,12] representing the topology of features present. Often, depth information is sparsely collected due to limitations of the depth sensors themselves.…”
Section: Related Workmentioning
confidence: 99%
“…Our choice of architecture is modeled from successful work in RGB-D object detection and classification utilizing feedforward neural networks ( Eitel et al, 2015 ; Ward et al, 2019 ). We utilize a dual-branch architecture for each input data modality along with feature map fusion ( Figure 2 ).…”
Section: Methodsmentioning
confidence: 99%
“…Our end-to-end approach eliminates the need to perform explicit individual plant segmentation and instead allows a deep convolutional neural network (DCNN) to implicitly perform segmentation by learning a mapping from input image space to individual plant biomass. Motivated by previous biomass estimation work that relies on 3D data as well as the ability of DCNNs to jointly learn from color and depth imagery ( Gupta et al, 2014 ; Eitel et al, 2015 ; Ophoff et al, 2019 ; Ward et al, 2019 ), our model incorporates both color and depth data as provided by an inexpensive and commercially available stereovision RGB-D camera. We hypothesize that DCNNs are well suited to understanding not only plant structure and size, but the influence of neighboring plant occlusion on the resulting view presented in overhead imagery of dense plant canopies.…”
Section: Introductionmentioning
confidence: 99%
“…Since 2010, when Microsoft, in cooperation with PrimeSense, released the first Kinect, consumer RGB-D cameras have been through a democratization process, becoming very appealing to many areas of application, such as robotics [ 1 , 2 ], automotive [ 3 ], industrial [ 4 ], augmented reality (AR) [ 5 ], object detection [ 6 ], 3D reconstruction [ 7 ], and biomedical field [ 8 ]. All these applications thrived by receiving depth information in addition to color.…”
Section: Introductionmentioning
confidence: 99%