2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2019
DOI: 10.1109/iros40897.2019.8967784
|View full text |Cite
|
Sign up to set email alerts
|

The CoSTAR Block Stacking Dataset: Learning with Workspace Constraints

Abstract: A robot can now grasp an object more effectively than ever before, but once it has the object what happens next? We show that a mild relaxation of the task and workspace constraints implicit in existing object grasping datasets can cause neural network based grasping algorithms to fail on even a simple block stacking task when executed under more realistic circumstances.To address this, we introduce the JHU CoSTAR Block Stacking Dataset (BSD), where a robot interacts with 5.1 cm colored blocks to complete an o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2

Relationship

3
3

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…In essence, the design steps of a neural network architecture that might otherwise be done by an engineer or graduate student by hand are instead automated and optimized as part of a well defined search space of reasonable layers, connections, outputs, and hyperparameters. In fact, architecture search can itself be defined in terms of hyperparameters [12] or as a graph search problem [27,19,2,24]. Furthermore, once a search space is defined various tools can be brought to bear on the problem including Bayesian optimization [16], other neural networks [1], reinforcement learning, evolution [21,20], or a wide variety of optimization frameworks.…”
Section: Related Workmentioning
confidence: 99%
“…In essence, the design steps of a neural network architecture that might otherwise be done by an engineer or graduate student by hand are instead automated and optimized as part of a well defined search space of reasonable layers, connections, outputs, and hyperparameters. In fact, architecture search can itself be defined in terms of hyperparameters [12] or as a graph search problem [27,19,2,24]. Furthermore, once a search space is defined various tools can be brought to bear on the problem including Bayesian optimization [16], other neural networks [1], reinforcement learning, evolution [21,20], or a wide variety of optimization frameworks.…”
Section: Related Workmentioning
confidence: 99%
“…Finally, neural architecture search forms the basis for our hyperparameter choices [23], [24]. Neural networks are imperfect arbitrary function approximators, so a better choice of algorithm is an effective approach to improving deep learning based robotic manipulation algorithms, as we have detailed in past work [25].…”
Section: Related Workmentioning
confidence: 99%
“…[bn, relu, conv1x1, bn, relu, conv1x1], where a 1x1 convolution is equivalent to a dense layer at each pixel. These parameters are based on the final dense block structure optimized for accuracy via HyperTree Architecture Search [25] in our prior work. We note that efficiency was not considered in the HyperTree metric and as a result this pixelwise dense block accounts for over 50% of the computation in EVT, so it is a good target for future efficiency gains.…”
Section: Action After Successful Grasp: Place (X Y !)mentioning
confidence: 99%
“…Motivated by a desire to enable better human-robot collaboration and finer-grained behavioral analyses, researchers in computer vision and robotics have recently begun to approach the challenging problem of assembly action recognition [1], [2], [3], [4], [5]. In assembly activity recognition, a perception system must recognize both the assembly actions and the configuration of a structure (e.g.…”
Section: Introductionmentioning
confidence: 99%