Deep reinforce learning for joint optimization of condition-based maintenance and spare ordering

Shengang Hao, Jun Zheng, Jie Yang, Haipeng Sun, Quanxin Zhang, Li Zhang*, Nan Jiang, Yuanzhang Li

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

13 Citations (Scopus)

Abstract

Condition-based maintenance (CBM) policy can avoid premature or late maintenance and reduce system failures and maintenance costs. Most existing CBM studies cannot solve the dimensional disaster problem in multi-component complex systems. Only some studies consider the constraint of maintenance resources when searching for the optimal maintenance policy, which is hard to apply to practical maintenance. This paper studies the joint optimization of the CBM policy and spare components inventory for the multi-component system in large state and action spaces. We use Markov Decision Process to model it and propose an improved deep reinforcement learning algorithm based on the stochastic policy and actor-critic framework. In this algorithm, factorization decomposes the system action into the linear combination of each component's action. The experimental results show that the algorithm proposed in this paper has better time performance and lower system cost compared with other benchmark algorithms. The training time of the former is only 28.5% and 9.12% of that of PPO and DQN algorithms, and the corresponding system cost is decreased by 17.39% and 27.95%, respectively. At the same time, our algorithm has good scalability and is suitable for solving Markov decision-making problems in large-scale state and action space.

Original languageEnglish
Pages (from-to)85-100
Number of pages16
JournalInformation Sciences
Volume634
DOIs
Publication statusPublished - Jul 2023

Keywords

  • Actor-critic framework
  • Condition-based maintenance
  • Deep reinforcement learning
  • Markov decision process
  • Stochastic policy

Fingerprint

Dive into the research topics of 'Deep reinforce learning for joint optimization of condition-based maintenance and spare ordering'. Together they form a unique fingerprint.

Cite this

Hao, S., Zheng, J., Yang, J., Sun, H., Zhang, Q., Zhang, L., Jiang, N., & Li, Y. (2023). Deep reinforce learning for joint optimization of condition-based maintenance and spare ordering. Information Sciences, 634, 85-100. https://doi.org/10.1016/j.ins.2023.03.064