2015 9th International Conference on Power Electronics and ECCE Asia (ICPE-ECCE Asia) 2015
DOI: 10.1109/icpe.2015.7167977
|View full text |Cite
|
Sign up to set email alerts
|

Q-learning algorithm based multi-agent coordinated control method for microgrids

Abstract: This paper proposes a Q-learning algorithm (Q-LA) based multi-agent coordinated control method for microgrids. By the method, Q-LA is adopted to calculate the power to be regulated, which is called the microgrid regulation error (MRE), in secondary control for real-time operation. And the generation schedule of distributed generators (DGs) as well as batteries is modified in real time with the MRE by the fuzzy theory and particle swarm optimization method, taking the economy and environmental benefits into con… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 13 publications
(12 reference statements)
0
4
0
Order By: Relevance
“…Noting that the action that gives the most reward is the greedy action, , strategies for solving the dilemma of exploration and exploitation have been developed, apart from random selection, namely; -greedy [31], the pursuit algorithm [64], the SoftMax function [30], random [65] and deterministic [45].…”
Section: B Exploration Exploitation Conundrummentioning
confidence: 99%
See 1 more Smart Citation
“…Noting that the action that gives the most reward is the greedy action, , strategies for solving the dilemma of exploration and exploitation have been developed, apart from random selection, namely; -greedy [31], the pursuit algorithm [64], the SoftMax function [30], random [65] and deterministic [45].…”
Section: B Exploration Exploitation Conundrummentioning
confidence: 99%
“…The algorithm converged to a policy that reduced energy interchange with the grid by 14%. Other applications of the use of Q-learning in energy scheduling in microgrids may be found in [65], [71] and [72].…”
Section: Q-learningmentioning
confidence: 99%
“…In the second scenarios, the BG2 is involved at 0.8 s and the first-level control will be activated at t = 1 s. The reactive power demand of DG 1 0 is decreased by 15 kVar at t = 1 s. It can be observed that the fluctuation of voltage can be eliminated because of the consensus protocol (12)- (14). The first-level control and the second-level control will not impact each other and the reactive power is decreased by 15 kVar at t = 2 s, t = 3 s to verify the validation.…”
Section: Performance Validationmentioning
confidence: 99%
“…The method based on Qlearning and graph theory is aimed to degrade the power grid pressure and increase the power grid stability. Q-learning is utilised for setting coordination control objectives and accelerating convergence characteristic [14]. On the other hand, Q-learning can optimise the power scheduling and increase the economy [15,16].…”
Section: Introductionmentioning
confidence: 99%