2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8794233
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Modal Geometric Learning for Grasping and Manipulation

Abstract: This work provides an architecture that incorporates depth and tactile information to create rich and accurate 3D models useful for robotic manipulation tasks. This is accomplished through the use of a 3D convolutional neural network (CNN). Offline, the network is provided with both depth and tactile information and trained to predict the object's geometry, thus filling in regions of occlusion. At runtime, the network is provided a partial view of an object. Tactile information is acquired to augment the captu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
54
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 49 publications
(54 citation statements)
references
References 36 publications
0
54
0
Order By: Relevance
“…To date, many grasping methods rely fully or in part on deep learning. Some methods only use deep learning to extract additional information about objects with e.g., shape completion [9], [10] or tactile information [11] and then use analytical methods to plan the actual grasp [12], while others employ data-driven grasp planning in an endto-end fashion to generate grasps directly from images [1]- [8]. We will review both shape completion and end-to-end data-driven grasp planning as both are vital parts of our grasping pipeline.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…To date, many grasping methods rely fully or in part on deep learning. Some methods only use deep learning to extract additional information about objects with e.g., shape completion [9], [10] or tactile information [11] and then use analytical methods to plan the actual grasp [12], while others employ data-driven grasp planning in an endto-end fashion to generate grasps directly from images [1]- [8]. We will review both shape completion and end-to-end data-driven grasp planning as both are vital parts of our grasping pipeline.…”
Section: Related Workmentioning
confidence: 99%
“…In the context of shape completion from incomplete pointclouds, most recent improvements come from the adoption of deep learning. For instance, different works have explored tailored network structures [9], [13], [14], semantic object classification to aid the reconstruction [15], the integration of other sensing modalities such as tactile information [11], or the exploitation of the network uncertainty [10].…”
Section: A Deep Shape Completionmentioning
confidence: 99%
See 1 more Smart Citation
“…Approaches to grasp synthesis can be classified into analytic and empirical methods; see Bohg et al [15] for a survey. Analytic approaches use physics-based contact models to compute force closure on an object, using the shape and estimated pose of the target object [16], [17], [18], but work poorly in the real world due to noisy sensing, simplified assumptions of contact physics, and difficulty in placing contact points accurately.…”
Section: B Grasp Synthesismentioning
confidence: 99%
“…Despite the fact that they successfully added context information to the voxel representations, their approach still required further improvements since the shape details of small 3D objects were missed. Later, Varley et al [3] presented a CNN approach for 3D shape reconstruction as part of a robot grasp planning algorithm from a single depth view [4]. This approach combined 3D convolutional layers with various fully connected layers to infer the complete 3D shape.…”
Section: Introductionmentioning
confidence: 99%