2019
DOI: 10.1016/j.apenergy.2019.01.140
|View full text |Cite
|
Sign up to set email alerts
|

Multi-agent reinforcement learning for modeling and control of thermostatically controlled loads

Abstract: Increasing energy efficiency of thermostatically controlled loads has the potential to substantially reduce domestic energy demand. However, optimizing the efficiency of thermostatically controlled loads requires either an existing model or detailed data from sensors to learn it online. Often, neither is practical because of real-world constraints. In this paper, we demonstrate that this problem can benefit greatly from multi-agent learning and collaboration. Starting with no thermostatically controlled load s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 69 publications
(28 citation statements)
references
References 34 publications
0
22
0
Order By: Relevance
“…Chen et al (2018) [66] benchmarked their controller with a "rule-based heuristic" control strategy. Kazmi et al (2019) [67] used a rule-based dead-band controller as the benchmark. Ahn and Park (2019) [68] claimed their controller saved 15.7% energy compared with the fixed pre-determined schedule on OA damper position and temperature setpoint.…”
Section: Performance Evaluationmentioning
confidence: 99%
“…Chen et al (2018) [66] benchmarked their controller with a "rule-based heuristic" control strategy. Kazmi et al (2019) [67] used a rule-based dead-band controller as the benchmark. Ahn and Park (2019) [68] claimed their controller saved 15.7% energy compared with the fixed pre-determined schedule on OA damper position and temperature setpoint.…”
Section: Performance Evaluationmentioning
confidence: 99%
“…Boltzmann probability distribution method [22] is used in this paper to describe the transition probability of the state in the evolutionary game. The Boltzmann probability distribution method selects the action by probability, and the probability of selecting the action a i in the state s is p(a i ) = e Q(s, a i )/λ ∑ a ∈ A e Q(s, a)/λ (16) where λ is the exponential function of the iteration period-k in evolutionary game. When λ increases, the agent's decision randomness also increases; and when λ decreases, the decision randomness also decreases.…”
Section: Regulation and Control Strategy On Decision Layer Based On Qmentioning
confidence: 99%
“…Moreover, in this paper, a MARL method is used for sequential decision making in multi-agent environment where traditional SARL is difficult to deal with. MARL has been adopted in some fields, such as vehicle routing problem [24] and thermostatically loads modeling [25]. The most universal MARL is equilibrium-based MARL, whose framework accords with Markov games and the evaluation of the learning process is based on all agents' joint behaviors, the equilibrium concept from game theory is introduced to denote optimal joint action [26][27][28][29][30].…”
Section: Related Workmentioning
confidence: 99%