2021
DOI: 10.1109/tie.2020.2988189
|View full text |Cite
|
Sign up to set email alerts
|

Online Recursive Power Management Strategy Based on the Reinforcement Learning Algorithm With Cosine Similarity and a Forgetting Factor

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 38 publications
(6 citation statements)
references
References 28 publications
0
5
0
Order By: Relevance
“…In [92], the SARSA algorithm was applied to address the issue of EMS for a FCHEV, the reward function was modeled as Gaussian distribution and the degree of hybridization was chosen as the action variable. A recursive algorithm was used to online update the TPM of demand power in [93],…”
Section: Rl-based Energy Management Strategiesmentioning
confidence: 99%
See 1 more Smart Citation
“…In [92], the SARSA algorithm was applied to address the issue of EMS for a FCHEV, the reward function was modeled as Gaussian distribution and the degree of hybridization was chosen as the action variable. A recursive algorithm was used to online update the TPM of demand power in [93],…”
Section: Rl-based Energy Management Strategiesmentioning
confidence: 99%
“…Q-learning [87][88][89][90][91], [93] Derive the optimal EMS for FCHEV, aim to improve the FCHEVs' performance SARSA [92] Compare the performance of Q-learning and SARSA in EMS for a FCHEV Q-learning [94,95] Propose an improve Q-learning, embed the recursive algorithm to update the TMP online Q-learning [96,97], [100] Combine the merits of Q-learning, PMP and DP Q-learning [98] Analyzes the impact of algorithm hyperparameters on EMS Policy iteration [99] Calculate the TPM of power demand, apply the EMS in real-time DP [101] Employ the DP in off-line training and ECMS in the on-line application Q-learning [102] Discuss the influence of the number of state variables in the Q-learning algorithm Dyna-H [103] Analyzes the difference between the Dyna-H and Q-learning…”
Section: Algorithms References Content Descriptionmentioning
confidence: 99%
“…Moreover, advanced multi-objective optimization methods [10]- [13] have been studied to mitigate the degradation of fuel cell and battery system, save fuel consumption, and keep battery charge sustaining. Recently, machine learning algorithms, such as online learning [14], reinforcement learning [15], [16], and rule learning algorithm [17], have been investigated for power management in an FCHEV system. However, the EMS using the optimal solution shows the result using simulation, and there are few real-time optimal EMS development results.…”
Section: Introductionmentioning
confidence: 99%
“…It is suitable for solving sequential decisionmaking problems. The purpose of reinforcement learning is to allow a reinforcement learning agent to learn how to behave in an environment where the only feedback consists of scalar learning signals, and the agent's goal is to maximize the reward signal from the environment in the long term [16,17]. Reinforcement learning algorithms capable of online parameter updates, fast convergence of the learning process, and suitable for different operating conditions have the potential to be applied to real-time energy management strategies [18][19][20].…”
Section: Introductionmentioning
confidence: 99%