Energy management based on reinforcement learning with double deep Q-learning for a hybrid electric tracked vehicle

Xuefeng Han, Hongwen He*, Jingda Wu, Jiankun Peng, Yuecheng Li

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

194 Citations (Scopus)

Abstract

An energy management strategy, based on double deep Q-learning algorithm, is proposed for a dual-motor driven hybrid electric tracked-vehicle. Typical model framework of tracked-vehicle is established where the lateral dynamic can be taken into consideration. For the propose of optimizing the fuel consumption performance, a double deep Q-learning-based control structure is put forward. Compared to conventional deep Q-learning, the proposed strategy prevents training process falling into the overoptimistic estimate of policy value and highlights its significant advantages in terms of the iterative convergence rate and optimization performance. Unique observation states are selected as input variables of reinforcement learning algorithm in view of revealing tracked-vehicles characteristic. The conventional deep Q-learning and dynamic programming are also employed and compared with the proposed strategy for different driving schedules. Simulation results demonstrate the fuel economy of proposed methodology achieves 7.1% better than that of conventional deep Q learning-based strategy and reaches 93.2% level of Dynamic programing benchmark. Moreover, the designed algorithm has a good performance in battery SOC retention with different initial values.

Original languageEnglish
Article number113708
JournalApplied Energy
Volume254
DOIs
Publication statusPublished - 15 Nov 2019

Keywords

  • Double deep Q-learning
  • Energy management
  • Hybrid vehicle
  • Reinforcement learning
  • Tracked-vehicle

Fingerprint

Dive into the research topics of 'Energy management based on reinforcement learning with double deep Q-learning for a hybrid electric tracked vehicle'. Together they form a unique fingerprint.

Cite this