IR 2021
DOI: 10.20517/ir.2021.02
|View full text |Cite
|
Sign up to set email alerts
|

Federated reinforcement learning: techniques, applications, and open challenges

Abstract: This paper presents a comprehensive survey of Federated Reinforcement Learning (FRL), an emerging and promising field in Reinforcement Learning (RL). Starting with a tutorial of Federated Learning (FL) and RL, we then focus on the introduction of FRL as a new method with great potential by leveraging the basic idea of FL to improve the performance of RL while preserving data-privacy. According to the distribution characteristics of the agents in the framework, FRL algorithms can be divided into two categories,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 59 publications
(8 citation statements)
references
References 75 publications
(105 reference statements)
0
8
0
Order By: Relevance
“…The TV penalty occurred when investigating the connection between the local and global policy advantage, i.e., Theorem I, to tackle the data heterogeneity, while the KL penalty came from (12) to constrain the policy update at one agent.…”
Section: Corollary I the Conditionmentioning
confidence: 99%
See 2 more Smart Citations
“…The TV penalty occurred when investigating the connection between the local and global policy advantage, i.e., Theorem I, to tackle the data heterogeneity, while the KL penalty came from (12) to constrain the policy update at one agent.…”
Section: Corollary I the Conditionmentioning
confidence: 99%
“…Due to the distributed data and computing power in large-scale applications such as autonomous driving, the training of RL algorithms under the FL framework is inevitable. Unfortunately, many challenges faced by supervised FL, e.g., the data heterogeneity and communication bottleneck, are still valid and even worse for FRL [12]. For example, centralized policy gradient methods already suffer from high variance which is detrimental to the convergence speed and training performance [13]- [15], and data heterogeneity imposes another layer of difficulty for the convergence of FRL.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…HFRL may be noted to have similarities to "Parallel RL". Parallel RL is a long studied field of RL, where agent gradients are transferred amongst each other [5,10,11] .…”
Section: Federated Reinforcement Learningmentioning
confidence: 99%
“…Still, many robots, especially those involved in tasks as consequential as search-and-rescue and space exploration [8], will have to adapt to ever-changing environmental conditions and continue to optimize and update their internal policies over the course of their lifetime [9]. As such, to untether these methods from the confines of a lab, data collection and storage must be memory efficient enough to allow for either low-latency networking [10], [11] or for sufficient experience data to be stored on-board edge computing devices [12]. As fast and secure updates may not be possible in remote locations or when using bandwidth-constrained or highlatency cloud networks [13], it is imperative to find ways to reduce the overall memory footprint of DRL training.…”
Section: Introductionmentioning
confidence: 99%