TY - JOUR
T1 - Energy management strategy optimization of fuel cell vehicles based on long-term and short-term hydrogen consumption prediction
AU - Hu, Donghai
AU - Xu, Yinjie
AU - Huang, Jixiang
AU - Lu, Dagang
AU - Wang, Jing
AU - Li, Jianwei
N1 - Publisher Copyright:
Copyright © 2025. Published by Elsevier Ltd.
PY - 2025/11
Y1 - 2025/11
N2 - Hydrogen consumption is not only an evaluation metric for the economic performance of fuel cell vehicles (FCVs), but also one of the key optimization objectives in energy management strategies (EMS). However, EMS that rely on cumulative hydrogen consumption (HC-C) or hydrogen consumption per 100 km (HC-P100) as optimization objectives are limited by the inaccuracies and time delays in hydrogen consumption data. This study proposes an algorithmic framework called Proximal Policy Optimization with Hydrogen Consumption Prediction (PPO-HCP) to optimize energy management strategies for FCVs. First, establishing a dynamic system model for FCVs. Then, innovatively designing a novel reward function to enhance the adaptability of deep reinforcement learning (DRL)-based EMS under complex and dynamic conditions. This includes weighting different terms in the reward function, such as short-term real-time hydrogen consumption (HC-RT), long-term HC-C and HC-P100, fuel cell power, and battery state of charge (SOC). Finally, the proposed PPO-HCP algorithm is evaluated and compared with the conventional PPO algorithm under both training and random conditions. The results show that the energy consumption optimization effect of the PPO-HCP algorithm is more significant, with HC-P100 reduced by 5.3 % under training conditions and 7.9 % under random conditions.
AB - Hydrogen consumption is not only an evaluation metric for the economic performance of fuel cell vehicles (FCVs), but also one of the key optimization objectives in energy management strategies (EMS). However, EMS that rely on cumulative hydrogen consumption (HC-C) or hydrogen consumption per 100 km (HC-P100) as optimization objectives are limited by the inaccuracies and time delays in hydrogen consumption data. This study proposes an algorithmic framework called Proximal Policy Optimization with Hydrogen Consumption Prediction (PPO-HCP) to optimize energy management strategies for FCVs. First, establishing a dynamic system model for FCVs. Then, innovatively designing a novel reward function to enhance the adaptability of deep reinforcement learning (DRL)-based EMS under complex and dynamic conditions. This includes weighting different terms in the reward function, such as short-term real-time hydrogen consumption (HC-RT), long-term HC-C and HC-P100, fuel cell power, and battery state of charge (SOC). Finally, the proposed PPO-HCP algorithm is evaluated and compared with the conventional PPO algorithm under both training and random conditions. The results show that the energy consumption optimization effect of the PPO-HCP algorithm is more significant, with HC-P100 reduced by 5.3 % under training conditions and 7.9 % under random conditions.
KW - Deep reinforcement learning
KW - Energy management strategy
KW - Fuel cell vehicle
KW - Hydrogen consumption
UR - https://www.scopus.com/pages/publications/105020982528
U2 - 10.1016/j.seta.2025.104650
DO - 10.1016/j.seta.2025.104650
M3 - Article
AN - SCOPUS:105020982528
SN - 2213-1388
VL - 83
JO - Sustainable Energy Technologies and Assessments
JF - Sustainable Energy Technologies and Assessments
M1 - 104650
ER -