Proceedings of the 33rd Chinese Control Conference 2014
DOI: 10.1109/chicc.2014.6896376
|View full text |Cite
|
Sign up to set email alerts
|

Optimal management of energy storage system based on reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 10 publications
0
3
0
Order By: Relevance
“…Using the "SARSA" algorithm, a control operation was developed to maximize long-term rewards to encounter the load demand. The results in [10] showed a significant increase of total rewards, a faster convergence speed and good effect. A reinforcement learning controller developed and simulated using the MATLAB/Simulink environment shows that as compared to the other controllers, the reinforcement learning controller has equivalent or better performance even after a couple of simulated years [11].…”
Section: Literature Reviewmentioning
confidence: 88%
“…Using the "SARSA" algorithm, a control operation was developed to maximize long-term rewards to encounter the load demand. The results in [10] showed a significant increase of total rewards, a faster convergence speed and good effect. A reinforcement learning controller developed and simulated using the MATLAB/Simulink environment shows that as compared to the other controllers, the reinforcement learning controller has equivalent or better performance even after a couple of simulated years [11].…”
Section: Literature Reviewmentioning
confidence: 88%
“…To solve (15), we can adopt certain RL algorithms to search for the optimal parameter θ; see e.g., [13]. We use the deep Qnetworks (DQNs) [20,Ch.…”
Section: Optimal Battery Control Algorithmmentioning
confidence: 99%
“…This data-driven framework helps to bypass the hurdles in formulating the complex models of system dynamics or estimating the statistical information on the uncertainty. Specifically for the OBC problem, it allows to flexibly incorporate a variety of operational objectives, and several RL techniques have been widely used, such as the Deep Q-Network (DQN) [14], SARSA [15] and T D(λ)-learning [16]. However, a majority of these techniques have not considered the battery degradation cost, as its cycle-based model is difficult to include by the RL formulation.…”
Section: Introductionmentioning
confidence: 99%