2016
DOI: 10.1177/1687814016668077
|View full text |Cite
|
Sign up to set email alerts
|

Robot grasp detection using multimodal deep convolutional neural networks

Abstract: Autonomous manipulation has enabled a wide range of exciting robot tasks. However, perceiving outside environment is still a challenging problem in the field of intelligent robotic research due to the lack of object models, unstructured environments, and time-consuming computation. In this article, we present a novel robot grasp detection system that maps a pair of RGB-D images of novel objects to best grasping pose of a robotic gripper. First, we segment the graspable objects from the unstructured scene using… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
86
0
2

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 132 publications
(92 citation statements)
references
References 30 publications
(41 reference statements)
0
86
0
2
Order By: Relevance
“…In their novel classification method for grasp detection, Zhou et al [34] used a similar five-element grasp rectangle representation following the previous work in [10,[18][19][20][31][32][33]35]. Wang et al [36] proposed a minor variation to this approach that differed simply by excluding the parameter for gripper plate height (h). They argued that this parameter can be controlled in the robotic set-up configurations thus the authors used a four-element grasp representation of G = (x, y, θ, w).…”
Section: Grasp Representationmentioning
confidence: 99%
See 2 more Smart Citations
“…In their novel classification method for grasp detection, Zhou et al [34] used a similar five-element grasp rectangle representation following the previous work in [10,[18][19][20][31][32][33]35]. Wang et al [36] proposed a minor variation to this approach that differed simply by excluding the parameter for gripper plate height (h). They argued that this parameter can be controlled in the robotic set-up configurations thus the authors used a four-element grasp representation of G = (x, y, θ, w).…”
Section: Grasp Representationmentioning
confidence: 99%
“…In the grasp detection work by Wang et al [36], the authors used the Washington RGB-D dataset [58] for its rich variety of RGB-D images. The authors self-annotated as they preferred to combine the resulting dataset with the CGD.…”
Section: Pre-compiled Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…A shallow network first predicts high-ranked candidate grasp rectangles, followed by a deeper network that chooses the optimal grasp points. Wang et al [38] followed a similar approach using a multi-modal CNN. Another method [15] uses RGB-D data to first extract features from a scene using a ResNet-50 architecture [11] and then a successive shallower convolutional network applied to the merged features to estimate the optimal point of grasping.…”
Section: Related Workmentioning
confidence: 99%
“…This tends to slow down the overall run-time and fails in presence of complicated or unseen object shapes. Following the success of deep learning in a wide spectrum of computer vision applications, several recent approaches [9,15,20,24,29,37,38] employed Convolutional Neural Networks (CNNs) [14,18] to successfully detect grasping points from visual data, typically parametrized by 5dimensional (5D) grasping representations [12,20]. It is worth noting that most of these methods rely on depth data, often paired with color information.…”
Section: Introductionmentioning
confidence: 99%