TY - JOUR
T1 - Deep reinforcement learning-based energy management for a series hybrid electric vehicle enabled by history cumulative trip information
AU - Li, Yuecheng
AU - He, Hongwen
AU - Peng, Jiankun
AU - Wang, Hong
N1 - Publisher Copyright:
© 2021 Georg Thieme Verlag. All rights reserved.
PY - 2019/8
Y1 - 2019/8
N2 - It is essential to develop proper energy management strategies (EMSs) with broad adaptability for hybrid electric vehicles (HEVs).This paper utilizes deep reinforcement learning (DRL) to develop EMSs for a series HEV due to DRL s advantages of requiring no future driving information in derivation and good generalization in solving energy management problem formulated as aMarkov decision process. History cumulative trip information is also integrated for effective state of charge guidance in DRLbased EMSs. The proposed method is systematically introduced from offline training to online applications; its learning ability, optimality, and generalization are validated by comparisons with fuel economy benchmark optimized by dynamic programming, and real-Time EMSs based on model predictive control (MPC). Simulation results indicate that without a priori knowledge of future trip, original DRL-based EMS achieves an average 3.5% gap from benchmark, superior to MPC-based EMS with accurate prediction; after further applying output frequency adjustment, a mean gap of 8.7%,which is comparablewithMPC-basedEMSwith mean prediction error of 1 m/s, is maintained with concurrently noteworthy improvement in reducing engine start times. Besides, its impressive computation speed of about 0.001 s per simulation step proves its practical application potential, and this method is independent of powertrain topology such that it is applicative for any type of HEVs even when future driving information is unavailable.
AB - It is essential to develop proper energy management strategies (EMSs) with broad adaptability for hybrid electric vehicles (HEVs).This paper utilizes deep reinforcement learning (DRL) to develop EMSs for a series HEV due to DRL s advantages of requiring no future driving information in derivation and good generalization in solving energy management problem formulated as aMarkov decision process. History cumulative trip information is also integrated for effective state of charge guidance in DRLbased EMSs. The proposed method is systematically introduced from offline training to online applications; its learning ability, optimality, and generalization are validated by comparisons with fuel economy benchmark optimized by dynamic programming, and real-Time EMSs based on model predictive control (MPC). Simulation results indicate that without a priori knowledge of future trip, original DRL-based EMS achieves an average 3.5% gap from benchmark, superior to MPC-based EMS with accurate prediction; after further applying output frequency adjustment, a mean gap of 8.7%,which is comparablewithMPC-basedEMSwith mean prediction error of 1 m/s, is maintained with concurrently noteworthy improvement in reducing engine start times. Besides, its impressive computation speed of about 0.001 s per simulation step proves its practical application potential, and this method is independent of powertrain topology such that it is applicative for any type of HEVs even when future driving information is unavailable.
KW - Deep reinforcement learning
KW - Dynamic programming
KW - Energy management
KW - Generalization
KW - Model predictive control
KW - Optimality.
UR - http://www.scopus.com/inward/record.url?scp=85080138165&partnerID=8YFLogxK
U2 - 10.1109/TVT.2019.2926472
DO - 10.1109/TVT.2019.2926472
M3 - Article
AN - SCOPUS:85080138165
SN - 0018-9545
VL - 68
SP - 7416
EP - 7430
JO - IEEE Transactions on Vehicular Technology
JF - IEEE Transactions on Vehicular Technology
IS - 8
M1 - 8754786
ER -