2020
DOI: 10.1109/jiot.2020.2988457
|View full text |Cite
|
Sign up to set email alerts
|

Stacked Autoencoder-Based Deep Reinforcement Learning for Online Resource Scheduling in Large-Scale MEC Networks

Abstract: An online resource scheduling framework is proposed for minimizing the sum of weighted task latency for all the mobile users, by optimizing offloading decision, transmission power, and resource allocation in the mobile edge computing (MEC) system. Towards this end, a deep reinforcement learning (DRL) method is proposed to obtain an online resource scheduling policy. Firstly, a related and regularized stacked auto encoder (2r-SAE) with unsupervised learning is proposed to perform data compression and representa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
33
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 44 publications
(36 citation statements)
references
References 37 publications
(40 reference statements)
0
33
0
Order By: Relevance
“…Table III characterizes the objective function values of the DIRS framework under different distributions of IoTDs. For evaluating the influence of the different position distributions of IoTDs, we define the normalized reward rate (NRR), which is equal to that the inferred reward dividing the optimal reward [44]. In NRR, the inferred reward in the numerator is calculated from the output of the DIRS framework, and the optimal reward in the denominator is obtained from the particle swarm optimization (PSO) algorithm which is always applied to solve large-scale MINLP problems with high quality but low efficiency [44].…”
Section: B Performance Evaluation For Different Modules Of Dirsmentioning
confidence: 99%
See 1 more Smart Citation
“…Table III characterizes the objective function values of the DIRS framework under different distributions of IoTDs. For evaluating the influence of the different position distributions of IoTDs, we define the normalized reward rate (NRR), which is equal to that the inferred reward dividing the optimal reward [44]. In NRR, the inferred reward in the numerator is calculated from the output of the DIRS framework, and the optimal reward in the denominator is obtained from the particle swarm optimization (PSO) algorithm which is always applied to solve large-scale MINLP problems with high quality but low efficiency [44].…”
Section: B Performance Evaluation For Different Modules Of Dirsmentioning
confidence: 99%
“…For evaluating the influence of the different position distributions of IoTDs, we define the normalized reward rate (NRR), which is equal to that the inferred reward dividing the optimal reward [44]. In NRR, the inferred reward in the numerator is calculated from the output of the DIRS framework, and the optimal reward in the denominator is obtained from the particle swarm optimization (PSO) algorithm which is always applied to solve large-scale MINLP problems with high quality but low efficiency [44]. From Table III, one can see that the IoTDs with nonuniform distributions (e.g., Gaussian distribution and Lévy distribution) obtain lower objective function values.…”
Section: B Performance Evaluation For Different Modules Of Dirsmentioning
confidence: 99%
“…Auto-encoder algorithm may be applied. For example, H-MEC could benefit from a pre-training scheme of Stacked Auto-Encoder (SAE) [12] for automatic feature learning. In particular, SAE can be applied to train an attack detection model with a mix of unlabelled normal/attack samples so that the model identifies patterns of attack and normal data by an auto-encoder scheme, this can in turn improve the accuracy of the attack detection model on unseen and mutated attacks.…”
Section: A Ai-based Solutionsmentioning
confidence: 99%
“…Simulated annealing (SA) is a probabilistic heuristic search technique based on the annealing process in metallurgy. Thanks to its fast convergence, less parameter, and simplicity, SA has been widely adapted for decision making and optimization in recent years [35]. In this paper, we propose an SA based gateway selection method, i.e., SAGA, to solve P2.1 and obtain the near optimal gateway selection decision, whose details are listed in Algorithm 2.…”
mentioning
confidence: 99%