2019 IEEE-RAS 19th International Conference on Humanoid Robots (Humanoids) 2019
DOI: 10.1109/humanoids43949.2019.9034991
|View full text |Cite
|
Sign up to set email alerts
|

Generative Adversarial Imitation Learning with Deep P-Network for Robotic Cloth Manipulation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(7 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…Building on top of Deep Deterministic Policy Gradients and Hindsight Experience Replay, Nair et al [272] proposed behavior cloning Loss to increase imitating the demonstrations. Besides Q-learning, Generative Adversarial Imitation Learning [364] proposes P-GAIL that integrates imitation learning into the policy gradient framework. P-GAIL considers both smoothness and causal entropy in policy update by utilizing Deep P-Network [365].…”
Section: Drl Recent Advancesmentioning
confidence: 99%
“…Building on top of Deep Deterministic Policy Gradients and Hindsight Experience Replay, Nair et al [272] proposed behavior cloning Loss to increase imitating the demonstrations. Besides Q-learning, Generative Adversarial Imitation Learning [364] proposes P-GAIL that integrates imitation learning into the policy gradient framework. P-GAIL considers both smoothness and causal entropy in policy update by utilizing Deep P-Network [365].…”
Section: Drl Recent Advancesmentioning
confidence: 99%
“…Specifically, during every iteration, online GAIL first minimizes the discrepancy in expected cumulative reward between the expert policy and the learned policy and then maximizes such a discrepancy over a given reward function class in adversary. Online GAIL achieves tremendous empirical success in a variety of fields, such as autonomous driving (Kuefler et al, 2017), human behavior modeling (Merel et al, 2017), natural language processing (Chen et al, 2017), and robotics control (Tsurumine et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…One such task, robotic cloth folding, is of particular interest to robotics community because of its high applicability. After early successes in robotic folding [4], [7], the field is now dominated by deep reinforcement learning approaches [8]- [10] where the most common approach is to perform the training in simulation before transferring the model to a real-world robot via domain randomization [9], [11]- [13]. Challenges for sim-to-real transfer of reinforcement learning policies are well-documented [14].…”
Section: Introductionmentioning
confidence: 99%
“…Sim-to-real is especially difficult for deformable objects because they can not be simulated accurately, and it is difficult to represent their state. Because of these challenges and encouraged by the successful use of expert human demonstrations in reinforcement learning [10], [15], research started exploring how to use demonstration data to train cloth folding policies [5], [11], [16], [17].…”
Section: Introductionmentioning
confidence: 99%