2018
DOI: 10.1109/access.2018.2876652
|View full text |Cite
|
Sign up to set email alerts
|

Local Energy Trading Behavior Modeling With Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
50
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 92 publications
(51 citation statements)
references
References 18 publications
0
50
0
Order By: Relevance
“…As an extension of Q-learning on multi-dimensional continuous state space, authors in [25], [26] proposed the deep Q network (DQN) method which employs a DNN to approximate the action-value function, and has performed at the level of expert humans in playing Atari 2600 games. Inspired by this pioneering work, several recent papers have employed the DQN method to various smart grid applications such as voltage control [27], residential load control [28], building energy management systems [29], electric vehicles [30] and energy storage scheduling [31] and energy trading for prosumers [32]. However, although previous work has demonstrated high quality performance of the DQN method in problems with continuous state spaces, its performance in problems with continuous action spaces is less satisfactory because the employed DNN is trained to produce discrete action-value estimates rather than continuous actions [33], which significantly hinders its effectiveness in addressing the examined market modeling problem, since market players' actions are continuous and multi-dimensional.…”
Section: A Background and Motivationmentioning
confidence: 99%
“…As an extension of Q-learning on multi-dimensional continuous state space, authors in [25], [26] proposed the deep Q network (DQN) method which employs a DNN to approximate the action-value function, and has performed at the level of expert humans in playing Atari 2600 games. Inspired by this pioneering work, several recent papers have employed the DQN method to various smart grid applications such as voltage control [27], residential load control [28], building energy management systems [29], electric vehicles [30] and energy storage scheduling [31] and energy trading for prosumers [32]. However, although previous work has demonstrated high quality performance of the DQN method in problems with continuous state spaces, its performance in problems with continuous action spaces is less satisfactory because the employed DNN is trained to produce discrete action-value estimates rather than continuous actions [33], which significantly hinders its effectiveness in addressing the examined market modeling problem, since market players' actions are continuous and multi-dimensional.…”
Section: A Background and Motivationmentioning
confidence: 99%
“…To cope with the emergency control in the power system, a novel adaptive strategy and an opensource platform were designed for the grid control [21]. Reference [25] proposed a deep Q-learning based algorithm for local energy trading to optimize the decision-making processes of prosumers. In [22], a novel two-timescale voltage regulation scheme was introduced for smart grids by coupling deep Q-learning with physics-based optimization.…”
Section: B Drl-based Power System Managementmentioning
confidence: 99%
“…Unlike traditional reinforcement learning, the DRL algorithms use powerful deep neuron networks to approximate their value function (such as Q-table), enabling automatic high-dimensional feature extraction and end-toend learning. Recently, the advantages of DRL were recognized by the community and some attempts were made to leverage DRL in various applications for electrical grid, including operational control [21]- [24], electricity market [25], [26], demand response [27] and energy management [28]. Although these applications presented advantageous results in their respective fields, several challenges were encountered.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…In [8], Chen et al proposed an innovative method based on the concept of 'energy trading'. Through an ad-hoc mathematical model of energy trading strategies of a prosumer in the proposed holistic market model, the prosumer's decision-making process will be analyzed as a Markov decision process so that the local market participation will be solved by deep reinforcement learning technology with experience replay mechanism.…”
Section: Deep Learning (Lstm) and Rl Based Trading Systems: Literaturmentioning
confidence: 99%