2019 IEEE Wireless Communications and Networking Conference Workshop (WCNCW) 2019
DOI: 10.1109/wcncw.2019.8902878
|View full text |Cite
|
Sign up to set email alerts
|

Joint Channel Selection and Power Control for NOMA: A Multi-Armed Bandit Approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(19 citation statements)
references
References 17 publications
0
16
0
Order By: Relevance
“…Controller selects the power allocation action a m t = (p m t , q m t ) according to Equations ( 20), (21).…”
Section: Multi-agent Td3-based Power Allocation (Mtpa) Networkmentioning
confidence: 99%
“…Controller selects the power allocation action a m t = (p m t , q m t ) according to Equations ( 20), (21).…”
Section: Multi-agent Td3-based Power Allocation (Mtpa) Networkmentioning
confidence: 99%
“…They formulate the problem as a noncooperative game and transform the nonconvex optimization problem into the convex form by using nonlinear fractional programming and solve the transformed problem by Dinkelbach’s method and Lagrangian duality theory [ 12 ]. Adjif et al adopted a multiarmed bandit-based method (MAB) in the uplink scenario with the goal of optimizing system throughput [ 13 ]. Zheng et al considered user selfishness, and modeled the NOMA uplink power allocation problem as a Nash bargaining game, which is solved by KKT Condition [ 14 ].…”
Section: Related Workmentioning
confidence: 99%
“…This allows one to essentially explore at the beginning of the learning and mostly to exploit the best arm found so far after a certain amount of time. The new exploration probability is defined as [ 22 , 31 ]: where is the exploration parameter. However, the main challenge of this policy is how to properly set the value of L .…”
Section: Power Allocation With Multi-armed Banditsmentioning
confidence: 99%
“…This allows one to essentially explore at the beginning of the learning and mostly to exploit the best arm found so far after a certain amount of time. The new exploration probability is defined as [22,31]:…”
Section: -Greedymentioning
confidence: 99%
See 1 more Smart Citation