2020
DOI: 10.14209/jcis.2020.7
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for QoS-Constrained Resource Allocation in Multiservice Networks

Abstract: In this article, we study a Radio Resource Allocation (RRA) that was formulated as a non-convex optimization problem whose main aim is to maximize the spectral efficiency subject to satisfaction guarantees in multiservice wireless systems. This problem has already been previously investigated in the literature and efficient heuristics have been proposed. However, in order to assess the performance of Machine Learning (ML) algorithms when solving optimization problems in the context of RRA, we revisit that prob… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 22 publications
(73 reference statements)
0
8
0
Order By: Relevance
“…The efficiency of this technique is mainly evaluated by its ability to minimize delay and energy consumption. Saraiva et al 40 developed a model using DRL for QoS constrained resource allocation in multiservice networks. This model offers improved performance in terms of throughput and outage rate.…”
Section: Related Workmentioning
confidence: 99%
“…The efficiency of this technique is mainly evaluated by its ability to minimize delay and energy consumption. Saraiva et al 40 developed a model using DRL for QoS constrained resource allocation in multiservice networks. This model offers improved performance in terms of throughput and outage rate.…”
Section: Related Workmentioning
confidence: 99%
“…Then, Adam optimisation is employed to update online network weights on the basis of F( ) i.e. loss function and defined as Adam optimisation is preferred over stochastic gradient descent algorithm in case of sparse gradient problems as it calculates independent adaptive learning rate (Saraiva et al 2020;Bhattacharya et al 2019) and efficiently reduces the loss function described in Eq. ( 26).…”
Section: Software-defined Wireless Networking Layermentioning
confidence: 99%
“…Therefore, the network state S t perceived through southbound interface is preprocessed to φ(S t ) and while every mapping, Q network generates a segment of memory comprising current state φ(S t ), current action a t , instant reward r t , and next state φ(S t+1 ) that are stacked in experience buffer D during the training process to improve the learning efficiency. In this way, the experience buffer strategy reduces the correlation among the training examples in order to avoid the divergence of optimal policy to the local minimum [27]. The choice of mini-batch size plays an important role in tuning deep learning systems.…”
Section: Agent Modulementioning
confidence: 99%
“…Adam optimisation algorithm has been adopted to update network weights iterative based on training data. In comparison to the stochastic gradient descent algorithm, Adam optimisation computes individual adaptive learning rates rather maintaining a single learning rate which improves the performance on sparse gradient problems and minimises mean square error E(θ ) [27,45]. The overall algorithm of agent module is divided into two parts, i.e.…”
Section: Agent Modulementioning
confidence: 99%
See 1 more Smart Citation