2020
DOI: 10.1109/lra.2020.3013937
|View full text |Cite
|
Sign up to set email alerts
|

Invariant Transform Experience Replay: Data Augmentation for Deep Reinforcement Learning

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 25 publications
(14 citation statements)
references
References 12 publications
0
12
0
Order By: Relevance
“…Therefore, it can be concluded that DDPG-NNFTSMC proposed in this paper has excellent control performance, and theoretically has good application prospects for industrial manipulators with uncertain dynamic models and general second-order nonlinear systems. Next, based on the research of scholars [12,31], we will try to apply DDPG-NNFTSMC to other different scenarios and combine it with deep reinforcement learning to further improve the control performance [32,36,37].…”
Section: Discussionmentioning
confidence: 99%
“…Therefore, it can be concluded that DDPG-NNFTSMC proposed in this paper has excellent control performance, and theoretically has good application prospects for industrial manipulators with uncertain dynamic models and general second-order nonlinear systems. Next, based on the research of scholars [12,31], we will try to apply DDPG-NNFTSMC to other different scenarios and combine it with deep reinforcement learning to further improve the control performance [32,36,37].…”
Section: Discussionmentioning
confidence: 99%
“…Interestingly, the value of approximate symmetries of MDPs has been explored in some theoretical work [45,50], but without architectures that can make use of it. Additionally, data augmentation, while not able to bake in architectural equivariance, has been successfully applied to encouraging equivariance on image tasks [30] and recently even on tabular state vectors [33,40].…”
Section: Related Workmentioning
confidence: 99%
“…Recent work on symmetries in single agent deep reinforcement learning has shown improvements in terms of data effiency. Such work revolves around symmetries in policy networks (van der Pol et al, 2020;Simm et al, 2020), symmetric filters (Clark & Storkey, 2014), invariant data augmentation (Laskin et al, 2020;Kostrikov et al, 2020) or equivariant trajectory augmentation (Lin et al, 2020;Mavalankar, 2020;Mishra et al, 2019) These approaches are only suitable for single agent problems or centralized multi-agent controllers. Here, we solve the problem of enforcing global equivariance while still allowing distributed execution.…”
Section: Related Workmentioning
confidence: 99%
“…Equivariant and geometric deep learning have gained traction in recent years, showing promising results in supervised learning (Cohen & Welling, 2016;Winkels & Cohen, 2018;Weiler et al, 2018;Weiler & Cesa, 2019;Worrall et al, 2017;Fuchs et al, 2020;Thomas et al, 2018), unsupervised learning (Dey et al, 2021) and reinforcement learning (van der Pol et al, 2020;Simm et al, 2020). In single agent reinforcement learning, enforcing equivariance to group symmetries has been shown to improve data efficiency, for example with MDP homomorphic networks (van der Pol et al, 2020), trajectory augmentation (Lin et al, 2020;Mavalankar, 2020), or symmetric locomotion policies (Abdolhosseini et al, 2019). Equivariant approaches enable a single agent to learn policies more efficiently within its environment by sharing weights between state-action pairs that are equivalent under a transformation.…”
Section: Introductionmentioning
confidence: 99%