2017
DOI: 10.1002/ett.3249
|View full text |Cite
|
Sign up to set email alerts
|

Learning‐based resource allocation in D2D communications with QoS and fairness considerations

Abstract: In device-to-device (D2D) communications, D2D users establish a direct link by utilizing the cellular users' spectrum to increase the network spectral efficiency. However, due to the higher priority of cellular users, interference imposed by D2D users to cellular ones should be controlled by channel and power allocation algorithms. Due to the unknown distribution of dynamic channel parameters, learning-based resource allocation algorithms work more efficient than classic optimization methods. In this paper, th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 36 publications
0
5
0
Order By: Relevance
“…Its autonomous and fast converging performance outperforms greedy, ϵ‐greedy Q‐learning and ranked heuristic‐based algorithm in terms of the average user throughput. Multi‐state Markovian decision process (MDP) is used to model RB allocation in U‐D2D in [18]. Recency‐based Q ‐value updating function is proposed to prevent the quick convergence to local optimal policy in conventional Q‐learning algorithm.…”
Section: Related Workmentioning
confidence: 99%
“…Its autonomous and fast converging performance outperforms greedy, ϵ‐greedy Q‐learning and ranked heuristic‐based algorithm in terms of the average user throughput. Multi‐state Markovian decision process (MDP) is used to model RB allocation in U‐D2D in [18]. Recency‐based Q ‐value updating function is proposed to prevent the quick convergence to local optimal policy in conventional Q‐learning algorithm.…”
Section: Related Workmentioning
confidence: 99%
“…Where 𝑔 𝑖 (𝑚) and 𝑝 𝑖 (𝑚) are the channel gain and transmission power at the slot 𝑚, respectively, and the white Gaussian noise power is shown as 𝜎 2 . At the denominator of Equation ( 4), channel gain and transmission powers are denoted as 𝑝 𝑗 (𝑚) and ℎ 𝑗 (𝑚) for the interfering D2D pairs.…”
Section: Queue Dynamicsmentioning
confidence: 99%
“…In the underlay model, D2D users and cellular users (CUs) can simultaneously share the uplink or downlink channels while in the overlay model there exist a dedicated resource for D2D either in uplink or downlink channel. In the literature, most of the studies considered underlay D2D communication due to its higher spectral gain over to overlay model [2]- [5]. In D2D communications that underlay the communication links of a cellular network, D2D users communicate directly with each other by sharing the radio resources with cellular users, either in a non-orthogonal or orthogonal manner [1].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Using the learning phenomenon, a strategy for allocation of resources to a D2D network is proposed in the work of Kazemi Rashed et al A low complexity algorithm has been proposed in the work of Cicalò and Tralli for maximizing the weighted sum rate. The number of D2D pairs and cellular users is bounded, which provides satisfactory service.…”
Section: Related Workmentioning
confidence: 99%