2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2019
DOI: 10.1109/iros40897.2019.8968142
|View full text |Cite
|
Sign up to set email alerts
|

Improving Robot Success Detection using Static Object Data

Abstract: We use static object data to improve success detection for stacking objects on and nesting objects in one another. Such actions are necessary for certain robotics tasks, e.g., clearing a dining table or packing a warehouse bin. However, using an RGB-D camera to detect success can be insufficient: same-colored objects can be difficult to differentiate, and reflective silverware cause noisy depth camera perception. We show that adding static data about the objects themselves improves the performance of an end-to… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
7
1

Relationship

2
6

Authors

Journals

citations
Cited by 12 publications
(11 citation statements)
references
References 46 publications
0
11
0
Order By: Relevance
“…As computers transition from desktops to pervasive mobile and edge devices, we must make and meet the expectation that NLP can be deployed in any of these contexts. Current representations have very limited utility in even the most basic robotic settings (Scalise et al, 2019), making collaborative robotics (Rosenthal et al, 2010) largely a domain of custom engineering rather than science.…”
Section: Ws4: Embodiment and Actionmentioning
confidence: 99%
“…As computers transition from desktops to pervasive mobile and edge devices, we must make and meet the expectation that NLP can be deployed in any of these contexts. Current representations have very limited utility in even the most basic robotic settings (Scalise et al, 2019), making collaborative robotics (Rosenthal et al, 2010) largely a domain of custom engineering rather than science.…”
Section: Ws4: Embodiment and Actionmentioning
confidence: 99%
“…Other works attempt to infer actions, rewards, or state-values of human videos and use them for learning predictive models [40] or RL [14,39]. Learning keypoint [51,8] or object/task centric representations from videos [42,38,34] is another promising strategy to learning rewards and representations between domains. Simulation has also been leveraged as supervision to learn such representations [32] or to produce human data with domain randomization [3].…”
Section: B Robotic Learning From Human Videosmentioning
confidence: 99%
“…The physical forces and sounds objects make during manipulation actions can also be associated with words such as rattling and heavy for multimodal understanding beyond vision [18,36]. Prior work has gathered language annotations for the YCB Benchmark object set [37] to explore how language descriptions provide priors on object affordances [38]. In our tabletop robot experiments, we use camera views of novel objects to evaluate zero shot transfer of LAGOR to the real world with minimal object rotations to achieve language-aligned camera views to select the correct referent object.…”
Section: Datamentioning
confidence: 99%

Language Grounding with 3D Objects

Thomason,
Shridhar,
Bisk
et al. 2021
Preprint
Self Cite