2022
DOI: 10.1049/tje2.12128
|View full text |Cite
|
Sign up to set email alerts
|

Deep Q‐network application for optimal energy management in a grid‐tied solar PV‐Battery microgrid

Abstract: This paper presents a deep Q‐network (DQN) technique to optimally manage energy resources in a microgrid in which the algorithm learns tasks in the same way as humans do. Every move the agent makes in the environment generates feedback which then motivate the agent to learn more about the environment and perform far more intelligent steps later in its learning stages. This paper proposes a DQN‐based energy management system that learns system uncertainties, including load demand, grid prices and volatile renew… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 49 publications
0
2
0
Order By: Relevance
“…Here, the operation constraints of the ESS system include the charging and discharging power constraint (16)(17)(18)(19), time period coupling constraint (20), battery capacity constraint (21), and inverter capacity constraint (22).…”
Section: 23mentioning
confidence: 99%
See 1 more Smart Citation
“…Here, the operation constraints of the ESS system include the charging and discharging power constraint (16)(17)(18)(19), time period coupling constraint (20), battery capacity constraint (21), and inverter capacity constraint (22).…”
Section: 23mentioning
confidence: 99%
“…DRL has both strong situational awareness and decision‐making ability and learns how to maximize rewards through continuous interaction with the environment in a trial‐and‐error manner, which provides solutions for complex physical control tasks [17]. In recent years, applications of DRL in MGs keep emerging, among which the commonly used algorithms are deep Q network [18] and its variants, such as multi‐agent deep Q network [19], double deep Q network [20], and deep deterministic policy gradient (DDPG) [21]. Though effective to some extent, DRL generally suffers from low data efficiency and time‐consuming training [22].…”
Section: Introductionmentioning
confidence: 99%