TY - JOUR
T1 - Reinforcement Learning of Adaptive Energy Management with Transition Probability for a Hybrid Electric Tracked Vehicle
AU - Liu, Teng
AU - Zou, Yuan
AU - Liu, Dexing
AU - Sun, Fengchun
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2015/12/1
Y1 - 2015/12/1
N2 - A reinforcement learning-based adaptive energy management (RLAEM) is proposed for a hybrid electric tracked vehicle (HETV) in this paper. A control oriented model of the HETV is first established, in which the state-of-charge (SOC) of battery and the speed of generator are the state variables, and the engine's torque is the control variable. Subsequently, a transition probability matrix is learned from a specific driving schedule of the HETV. The proposed RLAEM decides appropriate power split between the battery and engine-generator set (EGS) to minimize the fuel consumption over different driving schedules. With the RLAEM, not only is driver's power requirement guaranteed, but also the fuel economy is improved as well. Finally, the RLAEM is compared with the stochastic dynamic programming (SDP)-based energy management for different driving schedules. The simulation results demonstrate the adaptability, optimality, and learning ability of the RLAEM and its capacity of reducing the computation time.
AB - A reinforcement learning-based adaptive energy management (RLAEM) is proposed for a hybrid electric tracked vehicle (HETV) in this paper. A control oriented model of the HETV is first established, in which the state-of-charge (SOC) of battery and the speed of generator are the state variables, and the engine's torque is the control variable. Subsequently, a transition probability matrix is learned from a specific driving schedule of the HETV. The proposed RLAEM decides appropriate power split between the battery and engine-generator set (EGS) to minimize the fuel consumption over different driving schedules. With the RLAEM, not only is driver's power requirement guaranteed, but also the fuel economy is improved as well. Finally, the RLAEM is compared with the stochastic dynamic programming (SDP)-based energy management for different driving schedules. The simulation results demonstrate the adaptability, optimality, and learning ability of the RLAEM and its capacity of reducing the computation time.
KW - Adaptability
KW - Q-learning algorithm
KW - energy management
KW - hybrid electric tracked vehicle
KW - programming (SDP)
KW - state of charge (SOC)
KW - stochastic dynamic
UR - http://www.scopus.com/inward/record.url?scp=84947601121&partnerID=8YFLogxK
U2 - 10.1109/TIE.2015.2475419
DO - 10.1109/TIE.2015.2475419
M3 - Article
AN - SCOPUS:84947601121
SN - 0278-0046
VL - 62
SP - 7837
EP - 7846
JO - IEEE Transactions on Industrial Electronics
JF - IEEE Transactions on Industrial Electronics
IS - 12
M1 - 7234919
ER -