2019
DOI: 10.1016/j.apenergy.2019.113755
|View full text |Cite
|
Sign up to set email alerts
|

Multi-step reinforcement learning for model-free predictive energy management of an electrified off-highway vehicle

Abstract: Where a licence is displayed above, please note the terms and conditions of the licence govern your use of this document.When citing, please reference the published version. Take down policyWhile the University of Birmingham exercises care and attention in making items available there are rare occasions when an item has been uploaded in error or has been deemed to be commercially or otherwise sensitive.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
52
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
8

Relationship

2
6

Authors

Journals

citations
Cited by 116 publications
(52 citation statements)
references
References 37 publications
0
52
0
Order By: Relevance
“…The vehicle speed dataset is collected from the real driving conditions. To reduce the training time and improve the control accuracy, the vehicle velocity is divided into three-speed intervals that are [0-12] m/s, [12][13][14][15][16][17][18][19][20][21][22][23][24] m/s, [24][25][26][27][28][29][30][31][32][33][34][35][36] m/s, and they represent low, medium, and high speed, respectively. Then, the classified speed intervals are adopted to train the DDPG algorithm separately until the algorithm converges, the trained neural network is stored, as depicted in Fig.…”
Section: A a Bi-level Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…The vehicle speed dataset is collected from the real driving conditions. To reduce the training time and improve the control accuracy, the vehicle velocity is divided into three-speed intervals that are [0-12] m/s, [12][13][14][15][16][17][18][19][20][21][22][23][24] m/s, [24][25][26][27][28][29][30][31][32][33][34][35][36] m/s, and they represent low, medium, and high speed, respectively. Then, the classified speed intervals are adopted to train the DDPG algorithm separately until the algorithm converges, the trained neural network is stored, as depicted in Fig.…”
Section: A a Bi-level Frameworkmentioning
confidence: 99%
“…In [23], better power distributions between the battery and the ultracapacitor of PHEVs were obtained through the RL-based method, and 16.8% energy loss reduction was achieved. The researchers in [24] and [25] adopted the predicted EMS to improve the HEVs' performance. However, the discrete state space and action space of RL hinder its further application in EMS of HEVs, and the emergence of deep reinforcement learning (DRL) has bridged over this difficulty.…”
Section: Introductionmentioning
confidence: 99%
“…PHEVs are widely promoted as an efficient and clean solution that combines an internal combustion engine (ICE) with an electric motor and a large rechargeable battery. This hybrid powertrain enables all-electric driving for extended periods of time and overcomes the concern of range anxiety [5], [6]. Intensive efforts on PHEVs have developed energy management strategies (EMS) for coordinating the power split in a fuel-efficient way [7].…”
Section: Introductionmentioning
confidence: 99%
“…Liu et al implement Q-learning algorithms for energy management of series hybrid vehicles [32]. Zhou et al proposed multi-step reinforcement learning for energy management of a hybrid vehicle [33]. Cao et al optimize the energy use of a plug-in hybrid vehicle based on Q-learning [34].…”
Section: Introductionmentioning
confidence: 99%