2022
DOI: 10.1109/tcomm.2021.3135540
|View full text |Cite
|
Sign up to set email alerts
|

Decentralized Multi-Agent Power Control in Wireless Networks With Frequency Reuse

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 31 publications
0
6
0
Order By: Relevance
“…Within each framework, we highlight the motivations, the NN design principles, type of optimization problems, toy examples of its applications in 6G wireless networks, the theoretical analysis, the related research challenges as well as the summary of advantages and disadvantages. Specifically, the algorithm unrolling [14], learning to branch-and-bound (LBB) [15], GNN for structured optimization [22], deep reinforcement learning (DRL) for stochastic optimization [26], end-to-end learning for semantic optimization [27] as well as federated learning (FL) for distributed optimization [28] will be covered in the following sections to shed light on the excellent performance of ML compared with the conventional optimization algorithm in a variety of practical domains and provide guidance on the usage of ML techniques in 6G networks, followed by the summary of network design philosophies, theoretical tools, implementation issues and discussion of future directions to drive forward the research in this area. Before the detailed elaborations of specific MOA designs, we firstly summarize the existing design paradigms of MOAs from different perspectives in the following subsections.…”
Section: B Machine Learning In 6gmentioning
confidence: 99%
See 1 more Smart Citation
“…Within each framework, we highlight the motivations, the NN design principles, type of optimization problems, toy examples of its applications in 6G wireless networks, the theoretical analysis, the related research challenges as well as the summary of advantages and disadvantages. Specifically, the algorithm unrolling [14], learning to branch-and-bound (LBB) [15], GNN for structured optimization [22], deep reinforcement learning (DRL) for stochastic optimization [26], end-to-end learning for semantic optimization [27] as well as federated learning (FL) for distributed optimization [28] will be covered in the following sections to shed light on the excellent performance of ML compared with the conventional optimization algorithm in a variety of practical domains and provide guidance on the usage of ML techniques in 6G networks, followed by the summary of network design philosophies, theoretical tools, implementation issues and discussion of future directions to drive forward the research in this area. Before the detailed elaborations of specific MOA designs, we firstly summarize the existing design paradigms of MOAs from different perspectives in the following subsections.…”
Section: B Machine Learning In 6gmentioning
confidence: 99%
“…However, the above methods require additional CSI exchange between the neighboring BSs, which can downgrade the spectrum efficiency. To address this, DEC-MAPC proposed in [26] achieved fully decentralized power control only using local CSI while maximizing the sumrate of the network. Specifically, to achieve fully distributed implementation, DEC-MAPC was proposed to decompose the global state-action value into a monotonic increasing nonlinear function of all local state-action values.…”
Section: E Application 3: Distributed Constraint Optimization Problemsmentioning
confidence: 99%
“…, where d D k ,Bm is the distance between edge device D k and edge server B m and α is the path loss exponent. Based on the Jakes' fading model [54], small-scale block fading g D k ,Bm [s] is modeled as first-order complex Gauss-Markov process, i.e.,…”
Section: Communication Modelmentioning
confidence: 99%
“…The proposed DRL-based resource allocation algorithm can be directly applied when other correlated wireless channel models are considered. However, the uncorrelated channel (e.g., Rayleigh channel) will decrease the correlation of system state in two consecutive time, which may affect the estimation accuracy of the expected future reward and decrease the performance of resource allocation decisions [54].…”
Section: A Mdp Formulationmentioning
confidence: 99%
“…Especially in radio resource management, the allocation of resources is a difficult task. For instance, in [19], a deep reinforcement learning-based decentralized multiagent power control algorithm was proposed to improve the sum rate of a cellular network. In [20], multiagent deep reinforcement learning-based autonomous channel selection and transmission power selection were used to reduce the co-channel interference in a cellular network.…”
Section: Introductionmentioning
confidence: 99%