2018 IEEE-RAS 18th International Conference on Humanoid Robots (Humanoids) 2018
DOI: 10.1109/humanoids.2018.8625071
|View full text |Cite
|
Sign up to set email alerts
|

Visual Manipulation Relationship Network for Autonomous Robotics

Abstract: Robotic grasping is one of the most important fields in robotics and convolutional neural network (CNN) has made great progress in detecting robotic grasps. However, including multiple objects in one scene can invalidate the existing grasping detection algorithms based on CNN because of lacking of manipulation relationships among objects to guide the robot to grasp things in the right order. Therefore, the manipulation relationships are needed to help robot better grasp and manipulate objects. This paper prese… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
62
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 60 publications
(73 citation statements)
references
References 37 publications
1
62
0
Order By: Relevance
“…We can see that our model improves the performance of visual manipulation relationship reasoning compared with our previous work [6]. We assume that the improvements are achieved following these changes: (1) Different from our previous work, in this paper, we use ResNet-101 as the feature extractor, or called "backbone" instead of ResNet-50 and VGG-16 in [6]; (2) the object detector is Faster-RCNN from [18] instead of SSD in [19]; (3) the backbone is updated using multi-task loss function including grasp detection loss.…”
Section: Metricsmentioning
confidence: 58%
See 3 more Smart Citations
“…We can see that our model improves the performance of visual manipulation relationship reasoning compared with our previous work [6]. We assume that the improvements are achieved following these changes: (1) Different from our previous work, in this paper, we use ResNet-101 as the feature extractor, or called "backbone" instead of ResNet-50 and VGG-16 in [6]; (2) the object detector is Faster-RCNN from [18] instead of SSD in [19]; (3) the backbone is updated using multi-task loss function including grasp detection loss.…”
Section: Metricsmentioning
confidence: 58%
“…Recent works prove that CNNs achieve advanced performance on visual relationship reasoning [14]- [16]. Different from visual relationship, visual manipulation relationship [6] is proposed to solve the problem of grasping order in object stacking scenes with consideration of the safety and stability of objects. However, when this algorithm is directly combined with the grasp detection network to solve grasping problem in object stacking scenes, there are two main difficulties: 1) it is difficult to correctly match the detected grasps and the detected objects in object stacking scenes; 2) the cascade structure causes a lot of redundant calculations (e.g.…”
Section: B Visual Manipulation Relationship Reasoningmentioning
confidence: 99%
See 2 more Smart Citations
“…On the Cornell dataset, the accuracy of our scheme in image splitting and object splitting is 97.12% and 95.89%, respectively, which is equivalent to the most advanced grasping detection algorithm. In order to verify the effect of the scheme in multi-object scenarios, we used the VMRD dataset [ 14 ] containing multi-object scenarios, which produced an accuracy of 74.3%. As shown in Figure 1 , our method is also applied to real robot grasping tasks.…”
Section: Introductionmentioning
confidence: 99%