2018 IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR) 2018
DOI: 10.1109/simpar.2018.8376268
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning for non-prehensile manipulation: Transfer from simulation to physical system

Abstract: Reinforcement learning has emerged as a promising methodology for training robot controllers. However, most results have been limited to simulation due to the need for a large number of samples and the lack of automated-yet-safe data collection methods. Model-based reinforcement learning methods provide an avenue to circumvent these challenges, but the traditional concern has been the mismatch between the simulator and the real world. Here, we show that control policies learned in simulation can successfully t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
44
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 56 publications
(51 citation statements)
references
References 41 publications
1
44
0
Order By: Relevance
“…Combination of system identification and dynamics randomization has been used in the past to learn locomotion for a real quadruped [26], non-prehensile object manipulation [27] and in-hand object pivoting [28]. In our work, we recognize domain randomization and system identification as powerful tools for training general policies in simulation.…”
Section: Related Workmentioning
confidence: 99%
“…Combination of system identification and dynamics randomization has been used in the past to learn locomotion for a real quadruped [26], non-prehensile object manipulation [27] and in-hand object pivoting [28]. In our work, we recognize domain randomization and system identification as powerful tools for training general policies in simulation.…”
Section: Related Workmentioning
confidence: 99%
“…Manual parameter tuning is another form of simulator modification that can be done prior to applying reinforcement learning. Lowrey et al (2018) manually identify simulation parameters before applying policy gradient reinforcement learning to learn to push an object to target positions. Tan et al (2018) perform similar system identification (including disassembling the robot and making measurements of each part) and adding action latency modeling before using deep reinforcement learning to learn quadrapedal walking.…”
Section: Simulator Modificationmentioning
confidence: 99%
“…Domain randomization produces policies that are robust enough to transfer to the real world. An alternative approach that does not involve randomness is to learn policies that perform well under an ensemble of different simulators (Boeing & Bräunl, 2012;Rajeswaran et al, 2017;Lowrey et al, 2018). Pinto et al, (2017b) simultaneously learn an adversary that can perturb the learning agent's actions while it learns in simulation.…”
Section: Robustness Through Simulator Variancementioning
confidence: 99%
“…for v ∈ V do 10 P r .append(min-vertex(v,O)) 11 for v ∈ P r do 12 min d← ∞ 13 for p ∈ V c do 14 d← G(v, p) 15 if d<min d then 16 min d←d 17 if min d>max d then 18 max d←min d 19 return max d with the object, and the mesh of the desired contact region C d . To compute these metrics, we first project the desired contact region C d and the robot meshes L i∈N onto the object mesh O as shown in lines 5-10.…”
Section: Benchmark Guidelines a Scoringmentioning
confidence: 99%