A Deep Reinforcement Learning-Based Energy Management Framework with Lagrangian Relaxation for Plug-In Hybrid Electric Vehicle

Hailong Zhang, Jiankun Peng*, Huachun Tan*, Hanxuan Dong, Fan Ding

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

39 引用 (Scopus)

摘要

Reinforcement learning (RL)-based energy management is one of the current hot spots of hybrid electric vehicles. Recent advances in RL-based energy management focus on energy-saving performance but less considers the constrained setting for training safety. This article proposes an RL framework named coach-actor-double-critic (CADC) for the optimization of energy management considered as the constrained Markov decision process (CMDP). A bilevel onboard controller includes a neural network (NN)-based strategy actor and rule-based strategy coach for online energy management. Once the output of the actor exceeds the constrained range of feasible solutions, the coach would take charge of energy management to ensure safety. By using the Lagrangian relaxation, the optimization for CMDP transforms into an unconstrained dual problem to minimize the energy consumption while minimizing the coach participation. The parameters of the actor are updated in a manner of policy gradient through RL training with the Lagrangian value function. Double-critic with the same structure synchronously estimates the value function to avoid overestimate bias. Several experiments with the bus trajectories data demonstrate the optimality, self-learning ability, and adaptability of CADC. The results indicate that CADC outperforms the existing RL-based strategies and reaches above 95% energy-saving rate of the off-line global optimum.

源语言英语
文章编号9286514
页(从-至)1146-1160
页数15
期刊IEEE Transactions on Transportation Electrification
7
3
DOI
出版状态已出版 - 9月 2021
已对外发布

指纹

探究 'A Deep Reinforcement Learning-Based Energy Management Framework with Lagrangian Relaxation for Plug-In Hybrid Electric Vehicle' 的科研主题。它们共同构成独一无二的指纹。

引用此