2021
DOI: 10.1002/rnc.5516
|View full text |Cite
|
Sign up to set email alerts
|

A deep asynchronous actor‐critic learning‐based event‐triggered decentralized load frequency control of power systems with communication delays

Abstract: This article proposes a novel asynchronous advantage actor‐critic (A3C) learning‐based dynamic event‐triggered mechanism for the decentralized load frequency regulation to alleviate the local‐area communication burden and influence of the load fluctuations. The proposed dynamic event‐triggered mechanism applies the A3C algorithm to optimally adjust the threshold of the event‐triggered function in real time. In the A3C algorithm framework, the long short‐term memory (LSTM) network is used to estimate the policy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
11
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 47 publications
0
11
0
Order By: Relevance
“…The flow chart of the safe reward-based DRL control method for the EHSS constrained control is shown in Figure 2. R cbf in (11) is designed as the safe reward for the DDPG in the proposed control method, to implement the safety constraints of e s min ≤ e s ≤ e s max in the EHSS control. Then the optimal safety policy 𝜋 * D is learned for the DDPG with R cbf in the proposed control method.…”
Section: Algorithm For Safe Control Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…The flow chart of the safe reward-based DRL control method for the EHSS constrained control is shown in Figure 2. R cbf in (11) is designed as the safe reward for the DDPG in the proposed control method, to implement the safety constraints of e s min ≤ e s ≤ e s max in the EHSS control. Then the optimal safety policy 𝜋 * D is learned for the DDPG with R cbf in the proposed control method.…”
Section: Algorithm For Safe Control Methodsmentioning
confidence: 99%
“…In order to improve the safety of the optimal policy learned in the DRL for the nonlinear system with safety constraints while accelerating the convergence speed, a safe reward R cbf is shaped by combining R(x, u, x ′ ) with a CBF-based potential difference term R Γ as shown in (11). R Γ is designed in the form of the potential function according to the convergence guarantee of the reward shaping method, 24,25 to speed up the convergence process of the DRL control method.…”
Section: Safe Reward Shapingmentioning
confidence: 99%
See 3 more Smart Citations