2019
DOI: 10.48550/arxiv.1901.08277
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Federated Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
32
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(32 citation statements)
references
References 12 publications
0
32
0
Order By: Relevance
“…The participating BSs collaboratively train a model based on the local data contents to determine the optimal parameters for a global model. Similarly, the study in [86] utilises DRL under the FL setting to build a highquality model for agents taking into account their privacy. The proposed model is evaluated in two different domains, Text2Action and Grid-world.…”
Section: Network Slicing (Ns)mentioning
confidence: 99%
“…The participating BSs collaboratively train a model based on the local data contents to determine the optimal parameters for a global model. Similarly, the study in [86] utilises DRL under the FL setting to build a highquality model for agents taking into account their privacy. The proposed model is evaluated in two different domains, Text2Action and Grid-world.…”
Section: Network Slicing (Ns)mentioning
confidence: 99%
“…and number of A at T, n A (T ) increase in non-linear fashion as n(L i ) enlarges linearly. The upper bound of n A (T ) can be defined as (4).…”
Section: B Arithmetic Relation Through Dissecting Clustersmentioning
confidence: 99%
“…However, acquiring IID (Independently and Identically Distributed) data in a designated local machine is not guaranteed in a realworld distributed environment, as most of the existing systems tend to be biased and skewed due to intrinsic propensity and external factors that influence the circumstances. Thus, measures in order to upgrade the FL performance were introduced to successfully mitigate the unstable impact of aggregating non-IID data through approaches such as utilizing learning paradigms [4][5][6][7][8][9], clustering [10][11][12][13][14], knowledge distillation [15][16][17][18], stochastic approach [11], [19][20], Bayesian theorem [21][22][23][24], etc. Initially, FL architecture was specifically designed in order to accommodate the effective training among the distributed conditions without sending the local data itself for security purposes [1], [2], [25], thus many researches were conducted on FL to achieve higher privacy such as detecting adversaries [12], [26].…”
mentioning
confidence: 99%
“…[5] and [14] proposed federated RL algorithms for training policies to maximise content cache-hits on Base Stations at the network edge. [6] designed the FedRL system for training a policy, where individual FL clients do not have access to the full state-space of the RL task. While these works contribute to the development of model-free RL in the FL setting, they suffer from high sample complexity and lack theoretical guarantees.…”
Section: Related Workmentioning
confidence: 99%
“…More recently, federated Reinforcement Learning (RL) was proposed to extend FL to train RL models to solve sequential decision-making problems. The existing federated RL works [5], [6], [7] directly combine model-free RL with FL. Specifically, they train policies locally for all collaborating devices, using the model-free RL objective, and average the policy parameters on the server to generate a global policy for the next round of local training.…”
Section: Introductionmentioning
confidence: 99%