2019
DOI: 10.1109/tim.2018.2834085
|View full text |Cite
|
Sign up to set email alerts
|

RGB-DI Images and Full Convolution Neural Network-Based Outdoor Scene Understanding for Mobile Robots

Abstract: This paper presents a multisensor-based approach to outdoor scene understanding of mobile robots. Since laser scanning points in 3-D space are distributed irregularly and unbalanced, a projection algorithm is proposed to generate RGB, depth, and intensity (RGB-DI) images so that the outdoor environments can be optimally measured with a variable resolution. The 3-D semantic segmentation in RGB-DI cloud points is, therefore, transformed to the semantic segmentation in RGB-DI images. A full convolution neural net… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
10

Relationship

1
9

Authors

Journals

citations
Cited by 51 publications
(13 citation statements)
references
References 23 publications
0
13
0
Order By: Relevance
“…Furthermore, a multisensor-based approach using vision and laser sensors has been proposed in ref. [164] to generate features of the image such as RGB, depth and intensity (RGB-DI). A fully convolution network (FCN) with deep layers was designed to perform semantic segmentation of RGB-DI images.…”
Section: Deep Learningmentioning
confidence: 99%
“…Furthermore, a multisensor-based approach using vision and laser sensors has been proposed in ref. [164] to generate features of the image such as RGB, depth and intensity (RGB-DI). A fully convolution network (FCN) with deep layers was designed to perform semantic segmentation of RGB-DI images.…”
Section: Deep Learningmentioning
confidence: 99%
“…Another way to represent point cloud data in a 2D format is setting a virtual camera for collecting images from the scene. In addition to the RGB color and the depth that can be projected to the image, geometric features (e.g., normals, incidence angels) and radiometric information (e.g., intensity) can also be fed to the deep learning architecture (e.g., Zhuang, et al [ 171 ], Lawin, et al [ 172 ], Qiu, et al [ 173 ]). Instead of converting 3D data to 2D images, some other methods voxelize the point cloud data and develop deep learning techniques that can cope with voxels (e.g., Huang and You [ 174 ]).…”
Section: Classificationmentioning
confidence: 99%
“…In recent years, many methods of deep learning have been applied in practice. Qiu et al (2018) presented a multisensor-based approach to outdoor scene understanding of mobile robots, which uses fully convolutional neural network (FCN) to perform semantic segmentation for RGB, depth and intensity (RGB-DI) images. Levine et al (2016) proposed a method involving hand-eye coordination for robotic grasping from monocular images, which trained a large convolutional neural network to predict the probability of successful grasps by task-space motion of the gripper, using only monocular camera images and independently of camera calibration or the current robot pose.…”
Section: Introductionmentioning
confidence: 99%