2023
DOI: 10.1109/access.2023.3249572
|View full text |Cite
|
Sign up to set email alerts
|

Bridging the Reality Gap Between Virtual and Physical Environments Through Reinforcement Learning

Abstract: Creating Reinforcement learning(RL) agents that can perform tasks in the real-world robotic systems remains a challenging task due to inconsistencies between the virtual-and the real-world. This is known as the ''reality-gap'' which hinders the performance of a RL agent trained in a virtual environment. The research describes the techniques used to train the models, generate randomized environments, reward function, and techniques utilized to transfer the model to the physical environment for evaluation. For t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(11 citation statements)
references
References 33 publications
(31 reference statements)
0
4
0
Order By: Relevance
“…The Godot game engine provides a versatile and accessible platform that can be used to train DRL models. This framework was utilized in a recent study [29] to bridge the reality gap between virtual reality and reality using a 3-DoF Stewart platform. This research demonstrates the versatility and efficacy of using Godot game engines in DRL applications.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations
“…The Godot game engine provides a versatile and accessible platform that can be used to train DRL models. This framework was utilized in a recent study [29] to bridge the reality gap between virtual reality and reality using a 3-DoF Stewart platform. This research demonstrates the versatility and efficacy of using Godot game engines in DRL applications.…”
Section: Discussionmentioning
confidence: 99%
“…In this case, the environment is created in a game engine, which allows the randomization of attributes such as simulated fidelity, physics dynamics, lighting conditions, specular highlights, textures, and object positions and their orientation [22]. These augmentations in the virtual environment create complexity and diversity, allowing the agent to adjust its behavior progressively over time [29].…”
Section: Domain Randomizationmentioning
confidence: 99%
See 2 more Smart Citations
“…Additionally, Iriondo A et al employed the Twin Delayed Deep Deterministic Policy Gradient (TD3) method [17] to study the operation of picking up objects from a table using a mobile manipulator. Ranaweera M and colleagues enhanced training outcomes through domain randomization and the introduction of noise during the reinforcement learning process [18]. These methods share a core principle of incorporating probabilistic approaches to significantly reduce the impact of ineffective actions.…”
Section: Introductionmentioning
confidence: 99%