Reinforcement learning-based energy management strategy for a hybrid electric tracked vehicle

Research output: Contribution to journalArticlepeer-review

96 Citations (Scopus)

Abstract

This paper presents a reinforcement learning (RL)-based energy management strategy for a hybrid electric tracked vehicle. A control-oriented model of the powertrain and vehicle dynamics is first established. According to the sample information of the experimental driving schedule, statistical characteristics at various velocities are determined by extracting the transition probability matrix of the power request. Two RL-based algorithms, namely Q-learning and Dyna algorithms, are applied to generate optimal control solutions. The two algorithms are simulated on the same driving schedule, and the simulation results are compared to clarify the merits and demerits of these algorithms. Although the Q-learning algorithm is faster (3 h) than the Dyna algorithm (7 h), its fuel consumption is 1.7% higher than that of the Dyna algorithm. Furthermore, the Dyna algorithm registers approximately the same fuel consumption as the dynamic programming-based global optimal solution. The computational cost of the Dyna algorithm is substantially lower than that of the stochastic dynamic programming.

Original languageEnglish
Pages (from-to)7243-7260
Number of pages18
JournalEnergies
Volume8
Issue number7
DOIs
Publication statusPublished - 2015

Keywords

  • Dyna algorithm
  • Dynamic programming (DP)
  • Hybrid electric tracked vehicle (HETV)
  • Q-learning algorithm
  • Reinforcement learning (RL)
  • Stochastic dynamic programming (SDP)

Fingerprint

Dive into the research topics of 'Reinforcement learning-based energy management strategy for a hybrid electric tracked vehicle'. Together they form a unique fingerprint.

Cite this