2020
DOI: 10.1109/lra.2020.3015448
|View full text |Cite
|
Sign up to set email alerts
|

“Good Robot!”: Efficient Reinforcement Learning for Multi-Step Visual Tasks with Sim to Real Transfer

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
47
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 59 publications
(51 citation statements)
references
References 23 publications
1
47
1
Order By: Relevance
“…In practice, three elements are vital to help transfer learning improve its performance than building and training a network [31] from scratch: (i) Successful PTM can help the user remove hyper-parameter tuning; (ii) The initial layers in PTM can be thought of as feature descriptors, which extract low-level features, e.g., tints, edges, blobs, shades, and textures; (iii) The target model may only need to re-train the last several layers of the pre-trained model, since we believe the last several layers carry out the complex identification tasks. The basic idea of transfer learning is shown in Fig.…”
Section: Methodsmentioning
confidence: 99%
“…In practice, three elements are vital to help transfer learning improve its performance than building and training a network [31] from scratch: (i) Successful PTM can help the user remove hyper-parameter tuning; (ii) The initial layers in PTM can be thought of as feature descriptors, which extract low-level features, e.g., tints, edges, blobs, shades, and textures; (iii) The target model may only need to re-train the last several layers of the pre-trained model, since we believe the last several layers carry out the complex identification tasks. The basic idea of transfer learning is shown in Fig.…”
Section: Methodsmentioning
confidence: 99%
“…Gripping is also an important challenge in the robotic field, as performing robust gripping tasks in various situations is studied in many places [55], [56]. Unlike the above scenarios, gripping scenarios induce the combined contact situation between both dynamically moving hard and soft parts, which should be handled in a D-contact manner.…”
Section: Grippingmentioning
confidence: 99%
“…One of the approaches is focused on unscrewing operations in robotic disassembly of electronic waste using the Q-Learning method [16], other approaches have used a robotic arm to achieve a goal using the Deep Reinforcement Learning method DQN (Deep Q-Network) [5], TRPO(Trust Region Policy Optimization) [22] and tested the result of the experiment in a real application. Some approaches use 2D/3D cameras and some other sensors to observe the robotic environment [5,10,12], others use only dynamic simulation with the specified environment [13,16] or use real-time robot learning techniques [19].…”
Section: Related Workmentioning
confidence: 99%