2021
DOI: 10.48550/arxiv.2112.00359
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Tool as Embodiment for Recursive Manipulation

Abstract: Humans and many animals exhibit a robust capability to manipulate diverse objects, often directly with their bodies and sometimes indirectly with tools. Such flexibility is likely enabled by the fundamental consistency in underlying physics of object manipulation such as contacts and force closures. Inspired by viewing tools as extensions of our bodies, we present Tool-As-Embodiment (TAE), a parameterization for tool-based manipulation policies that treat hand-object and tool-object interactions in the same re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 50 publications
0
1
0
Order By: Relevance
“…Arguably, a fundamental bottleneck for pretraining in RL is the difficulty in reusing a single network across vastly different tasks, of distinct observation spaces, action spaces, rewards, scenes, and agent morphologies. Preliminary work explored various aspects of this problem through graph neural networks for morphology generalization (Wang et al, 2018b;Pathak et al, 2019;Chen et al, 2018;Kurin et al, 2020), language for universal reward specification (Jiang et al, 2019;Lynch & Sermanet, 2021;Shridhar et al, 2022), and object-centric action spaces (Zeng et al, 2020;Shridhar et al, 2022;Noguchi et al, 2021). Our work is orthogonal to these as we essentially amortize RL algorithm itself, expressed as sequence modeling with Transformer, instead of specific RL domain information, and can be combined with domain-specific pre-training techniques (Yen- Chen et al, 2020;Lynch & Sermanet, 2021) effortlessly.…”
Section: Ablation Of Proposed Techniquesmentioning
confidence: 99%
“…Arguably, a fundamental bottleneck for pretraining in RL is the difficulty in reusing a single network across vastly different tasks, of distinct observation spaces, action spaces, rewards, scenes, and agent morphologies. Preliminary work explored various aspects of this problem through graph neural networks for morphology generalization (Wang et al, 2018b;Pathak et al, 2019;Chen et al, 2018;Kurin et al, 2020), language for universal reward specification (Jiang et al, 2019;Lynch & Sermanet, 2021;Shridhar et al, 2022), and object-centric action spaces (Zeng et al, 2020;Shridhar et al, 2022;Noguchi et al, 2021). Our work is orthogonal to these as we essentially amortize RL algorithm itself, expressed as sequence modeling with Transformer, instead of specific RL domain information, and can be combined with domain-specific pre-training techniques (Yen- Chen et al, 2020;Lynch & Sermanet, 2021) effortlessly.…”
Section: Ablation Of Proposed Techniquesmentioning
confidence: 99%