2021
DOI: 10.1155/2021/6629852
|View full text |Cite
|
Sign up to set email alerts
|

DDPG‐Based Energy‐Efficient Flow Scheduling Algorithm in Software‐Defined Data Centers

Abstract: With the rapid development of data centers, the energy consumption brought by more and more data centers cannot be underestimated. How to intelligently manage software-defined data center networks to reduce network energy consumption and improve network performance is becoming an important research subject. In this paper, for the flows with deadline requirements, we study how to design the rate-variable flow scheduling scheme to realize energy-saving and minimize the mean completion time (MCT) of flows based o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 24 publications
0
4
0
Order By: Relevance
“…It uses the offline learning method and Q-network for training and takes samples from the replay buffer to minimize correlation between samples. The authors in [22] adapt the DDPG algorithm to find the optimal scheduling scheme for flows. The authors in [23,24] present a QoS optimization algorithm based on DDPG that ultimately improves the load-balancing degree and throughput rate to ensure delay and packet-loss rate.…”
Section: Related Researchmentioning
confidence: 99%
See 1 more Smart Citation
“…It uses the offline learning method and Q-network for training and takes samples from the replay buffer to minimize correlation between samples. The authors in [22] adapt the DDPG algorithm to find the optimal scheduling scheme for flows. The authors in [23,24] present a QoS optimization algorithm based on DDPG that ultimately improves the load-balancing degree and throughput rate to ensure delay and packet-loss rate.…”
Section: Related Researchmentioning
confidence: 99%
“…Reinforcement learning opens a new way for solving complex network problems [15]. Some researchers have used traditional algorithms of reinforcement learning such as deep Q-learning network (DQN) [16][17][18][19], proximal policy optimization (PPO) [20], deep deterministic policy gradient (DDPG) [21][22][23][24][25][26][27][28], and twin delayed deep deterministic policy gradient (TD3) [29][30][31]. The DQN algorithm uses Q-tables to store value functions, but it leads to excessive memory overhead and sizeable computational complexity when the network size increases.…”
Section: Introductionmentioning
confidence: 99%
“…Equation ( 5) has already been proved for its establishment [29,30]. Even the zero mean error of the initial state will lead to an overestimation of the action value due to the update of the value function, and the adverse efect of this error will be gradually enlarged by the calculation of the Bellman equation.…”
Section: Error Analysis It Is An Inevitable Problem For Q-mentioning
confidence: 99%
“…For the dynamic and stochastic nature of order dispatching in ride-sharing platforms, Tang X. et al [24] proposed an order dispatching solution based on deep reinforcement learning, and verified the effectiveness of the algorithm through large-scale online tests. In addition, the application of DRL to network flow control problems [25], financial market intraday trading [26], subway train dispatching [27], etc. proved the superiority and effectiveness of DRL in solving sequence decision-making.…”
Section: Literature Reviewmentioning
confidence: 99%