2020 IEEE International Conference on Robotics and Automation (ICRA) 2020
DOI: 10.1109/icra40945.2020.9197409
|View full text |Cite
|
Sign up to set email alerts
|

Accurate Vision-based Manipulation through Contact Reasoning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(4 citation statements)
references
References 21 publications
0
4
0
Order By: Relevance
“…Past work in pushing manipulation incorporates simulation and analytical models in a motion planning loop to get the next feasible state [13] [14]. Additional contributions in push modeling involve combining object state estimation with affordance prediction from image data to determine contact points for achieving the optimal push [15] and creating a deep recurrent neural network model to model push outcomes for a variety of objects [16]. However, both approaches use a greedy planner operating in obstaclefree environments.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Past work in pushing manipulation incorporates simulation and analytical models in a motion planning loop to get the next feasible state [13] [14]. Additional contributions in push modeling involve combining object state estimation with affordance prediction from image data to determine contact points for achieving the optimal push [15] and creating a deep recurrent neural network model to model push outcomes for a variety of objects [16]. However, both approaches use a greedy planner operating in obstaclefree environments.…”
Section: Background and Related Workmentioning
confidence: 99%
“…In Bejjani et al (2019), rendered images from a simulator were used as state representation to exploit the generalization ability of CNNs. The work of Kloss et al (2020) exploits imagebased representations for pushing scenarios, similar to our encoding. In our work, the network learns a representation in image space that is able to reason over complex action sequences from an initial observation only and is able to generalize over changing numbers of objects.…”
Section: Learning Heuristics For Tamp and Mip In Roboticsmentioning
confidence: 99%
“…Prior work in visual manipulation has involved learning predictive models from videos of a robot interacting with its environment. These models are used to find actions with simulation roll-outs, in a model-predictive fashion [16,17,1,18,19,20,21]. These approaches have been successful at solving prehensile manipulation problems where the dynamics are hard to model.…”
Section: Related Workmentioning
confidence: 99%