2019
DOI: 10.1016/j.apenergy.2018.11.002
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning for demand response: A review of algorithms and modeling techniques

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
184
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 520 publications
(227 citation statements)
references
References 99 publications
1
184
0
Order By: Relevance
“…Several studies in the literature have also considered DSM schemes that involve heating, ventilation, and air conditioning (HVAC) systems, as it is one of the dominating electricity consuming systems in buildings [33][34][35]. Although HVAC systems are associated with occupants' comfort [36], different DSM strategies are proposed which do not breach thermal comfort of the occupants. In addition, impacts on comfort also depend on thermal characteristics of the buildings [37].…”
Section: Dsm In Developed and Developing Countriesmentioning
confidence: 99%
See 3 more Smart Citations
“…Several studies in the literature have also considered DSM schemes that involve heating, ventilation, and air conditioning (HVAC) systems, as it is one of the dominating electricity consuming systems in buildings [33][34][35]. Although HVAC systems are associated with occupants' comfort [36], different DSM strategies are proposed which do not breach thermal comfort of the occupants. In addition, impacts on comfort also depend on thermal characteristics of the buildings [37].…”
Section: Dsm In Developed and Developing Countriesmentioning
confidence: 99%
“…Numerous studies have explained 'non-stationary' or 'dynamic' demand response using different algorithms [12,36,[77][78][79][80]. Most of these studies are simulation based.…”
Section: Types Of Demand Response and Its Potentialitymentioning
confidence: 99%
See 2 more Smart Citations
“…RL methods can solve sequential decisionmaking problems in real time [10]. The last two decades have seen increasing efforts to apply conventional RL methods, such as Q-learning and fitted Q-iteration [10], in various decisionmaking and control problems in power systems; these range from demand response [11], energy management, and automatic generation control to transient stability and emergency control [9], [12], [13]. Due to scalability issues, applications of conventional RL methods are mainly focusing on problems with low-dimensional state and action spaces.…”
Section: Introductionmentioning
confidence: 99%