2021
DOI: 10.3390/s21061960
|View full text |Cite
|
Sign up to set email alerts
|

Deep Q-Learning for Two-Hop Communications of Drone Base Stations

Abstract: In this paper, we address the application of the flying Drone Base Stations (DBS) in order to improve the network performance. Given the high degrees of freedom of a DBS, it can change its position and adapt its trajectory according to the users movements and the target environment. A two-hop communication model, between an end-user and a macrocell through a DBS, is studied in this work. We propose Q-learning and Deep Q-learning based solutions to optimize the drone’s trajectory. Simulation results show that, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 31 publications
(30 reference statements)
0
6
0
Order By: Relevance
“…IAB has been justified for usage over 5G infrastructure by the 3GPP [114] and is deemed as useful in enhancing capacity, coverage, as well as connectivity. However, additional challenges are imposed on the UAV that needs to guarantee stable backhaul and access links [115]. Cao et al in [116] proposed a UEdriven deep reinforcement learning (DRL) based scheme, in which a centralized agent deployed at the backhaul side of NT-BSs is responsible for training the parameter of a deep Qnetwork (DQN), and each UE is able to access a proper NT-BS intelligently to enhance the long-term system throughput and avoid frequent handovers among NT-BSs.…”
Section: B Integrated Access and Backhaulmentioning
confidence: 99%
“…IAB has been justified for usage over 5G infrastructure by the 3GPP [114] and is deemed as useful in enhancing capacity, coverage, as well as connectivity. However, additional challenges are imposed on the UAV that needs to guarantee stable backhaul and access links [115]. Cao et al in [116] proposed a UEdriven deep reinforcement learning (DRL) based scheme, in which a centralized agent deployed at the backhaul side of NT-BSs is responsible for training the parameter of a deep Qnetwork (DQN), and each UE is able to access a proper NT-BS intelligently to enhance the long-term system throughput and avoid frequent handovers among NT-BSs.…”
Section: B Integrated Access and Backhaulmentioning
confidence: 99%
“…Returning to Equation ( 9), x v is given by Equation ( 14), and the 3-dB beamwidth in the elevation plane θ 3 is calculated using Equation (15), where G 0 represents the maximum gain in the azimuth plane. In addition, x k is calculated using Equation ( 16), where k v is an elevation pattern adjustment factor based on the leaked power.…”
Section: Interfered System Modelmentioning
confidence: 99%
“…In this regard, local dithering techniques like ϵgreedy algorithm is commonly used to allow large exploration (with probability ϵ) at the beginning of the training phase and later increases exploitation with increasing number of training episodes. In recent studies, [38]- [40] DQN is often used with the aforementioned tricks and trades (experienced replay, target network with ϵ-greedy algorithm). Collectively, we refer to them as vanilla DQN for the rest of the article.…”
Section: A Reinforcement Learning and Vanilla Dqnmentioning
confidence: 99%