2018 International Conference on Advanced Systems and Electric Technologies (IC_ASET) 2018
DOI: 10.1109/aset.2018.8379860
|View full text |Cite
|
Sign up to set email alerts
|

An adaptive Q-learning approach to power control for D2D communications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 12 publications
(12 citation statements)
references
References 6 publications
0
12
0
Order By: Relevance
“…where f i (e) is defined in ( 12) and e i represents the service order of user i. The problem (13) aims to find the optimal trajectory so that the UAV can complete most of the users' requests within their endurance time after receiving all the user requests. Since finding the optimal trajectory needs to evaluate all possible permutations of service order e, which takes up a substantial amount of service time, it's essential to introduce a learning algorithm to shorten the calculation time for the trajectory.…”
Section: Problem Formulationmentioning
confidence: 99%
See 2 more Smart Citations
“…where f i (e) is defined in ( 12) and e i represents the service order of user i. The problem (13) aims to find the optimal trajectory so that the UAV can complete most of the users' requests within their endurance time after receiving all the user requests. Since finding the optimal trajectory needs to evaluate all possible permutations of service order e, which takes up a substantial amount of service time, it's essential to introduce a learning algorithm to shorten the calculation time for the trajectory.…”
Section: Problem Formulationmentioning
confidence: 99%
“…To solve the maximization problem in (13), we introduce a reinforcement learning framework based on double Q-learning. Compared to the existing reinforcement learning algorithms [12]- [14] such as Q-learning that may result in sub-optimal trajectory and leads to the number of satisfied users not maximized, the proposed double Q-learning algorithm enables the UAV to find the optimal flying trajectory to serve the users so as to maximize the number of satisfied users.…”
Section: Double Q-learning Framework For Maximizing the Number Of Sat...mentioning
confidence: 99%
See 1 more Smart Citation
“…The reinforcement learning (RL) based resource allocation schemes have been applied to device-to-device (D2D) communication widely. In [26], a Q-learning based power control algorithm was proposed which decorrelated the actions selected by users and expand the solution space, and it had higher quality of service (QoS) than the schemes based on correlated Q-learning. In [27], two RL based power control methods were proposed, i.e., centralized method and distributed method.…”
Section: Introductionmentioning
confidence: 99%
“…Solutions are exploring on low-complexity methods to small-cell base-station design appropriate for future 5G indoor deployments. 21 Machine learning is one of the tools that could provide best set of solutions to learn the influential scenarios and certain parameters of the communication networks. The reinforcement leaning 11 in machine learning holds much power over the D2D communication network due to its self-healing nature using many corrective actions.…”
Section: Introductionmentioning
confidence: 99%