2022
DOI: 10.1016/j.egyr.2022.02.231
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic economic dispatch of power system based on DDPG algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 6 publications
0
7
0
Order By: Relevance
“…As will be explained later, reinforcement learning is a suitable technique that could be applied to power plant operation, since it is based on rewarding desired actions and penalizes undesired ones so that the system learns autonomously the optimal action to be taken without running an optimizer each time of its operation [ 25 ]. In this context, Liu, Liu [ 26 ] presented a Deep Deterministic Policy Gradient (DDPG) model (which is one of the reinforcement learning algorithms that relies on neural networks) combining conventional generators, wind turbines, and solar PVs. The model aims to minimize the costs of power generation coming from the deviation from forecasted power penalties and the cost of fuel.…”
Section: Literature Reviewmentioning
confidence: 99%
See 1 more Smart Citation
“…As will be explained later, reinforcement learning is a suitable technique that could be applied to power plant operation, since it is based on rewarding desired actions and penalizes undesired ones so that the system learns autonomously the optimal action to be taken without running an optimizer each time of its operation [ 25 ]. In this context, Liu, Liu [ 26 ] presented a Deep Deterministic Policy Gradient (DDPG) model (which is one of the reinforcement learning algorithms that relies on neural networks) combining conventional generators, wind turbines, and solar PVs. The model aims to minimize the costs of power generation coming from the deviation from forecasted power penalties and the cost of fuel.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The model aims to minimize the costs of power generation coming from the deviation from forecasted power penalties and the cost of fuel. When compared to a model predictive control approach, Liu, Liu [ 26 ] found that DDPP yields lower uncertainty cost, meaning a lower deviation from the forecast. However, the method is not compared with optimization and did not include heat supply which may complicate the modeling.…”
Section: Literature Reviewmentioning
confidence: 99%
“…16 The DDPG algorithm employs an Actor-Critic network to approximate the policy function 𝜇 and utilizes the DQN algorithm to train the network function Q, which enables the computation of temporal difference errors and the implementation of gradient updates from the Online Network to the Target Network. 17 Q function in the Critic network represents the expected value R t obtained after executing action a t output by actor network and policy 𝜇 in state s t , with a discount factor of 𝛾:…”
Section: Ddpg-based Control Algorithmmentioning
confidence: 99%
“…Deep deterministic policy gradient was proposed by the DeepMind team in 2016 as a strategy algorithm that incorporates deep learning neural networks into DPG 16 . The DDPG algorithm employs an Actor‐Critic network to approximate the policy function μ$$ \mu $$ and utilizes the DQN algorithm to train the network function Q, which enables the computation of temporal difference errors and the implementation of gradient updates from the Online Network to the Target Network 17 …”
Section: Ddpg‐based Control Algorithmmentioning
confidence: 99%
“…Zhang et al proposed a deep-reinforcement-learning-based energy scheduling strategy to optimize multiple targets, taking diversified uncertainties into account; an integrated power, heat, and natural gas system consisting of energy-coupling units and wind power generation interconnected via a power grid was modeled as a Markov decision process [ 35 ]. Liu et al proposed an adaptive uncertain dynamic economic dispatch method based on deep deterministic policy gradient (DDPG); on the basis of the economic dispatch model, they built a Markov decision process for power systems [ 36 ]. In this paper, the operation optimization of the sugarcane milling process is described as an MDP process, which is modeled as follows:…”
Section: Solving the Collaborative Optimization Model Of Mf-ef-if In ...mentioning
confidence: 99%