Optimizing Constrained Guidance Policy with Minimum Overload Regularization

Weilin Luo, Lei Chen, Kexin Liu, Haibo Gu, Jinhu Lu*

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

3 引用 (Scopus)

摘要

Using reinforcement learning (RL) algorithm to optimize guidance law can address non-idealities in complex environment. However, the optimization is difficult due to huge state-Action space, unstable training, and high requirements on expertise. In this paper, the constrained guidance policy of a neural guidance system is optimized using improved RL algorithm, which is motivated by the idea of traditional model-based guidance method. A novel optimization objective with minimum overload regularization is developed to restrain the guidance policy directly from generating redundant missile maneuver. Moreover, a bi-level curriculum learning is designed to facilitate the policy optimization. Experiment results show that the proposed minimum overload regularization can reduce the vertical overloads of missile significantly, and the bi-level curriculum learning can further accelerate the optimization of guidance policy.

源语言英语
页(从-至)2994-3005
页数12
期刊IEEE Transactions on Circuits and Systems I: Regular Papers
69
7
DOI
出版状态已出版 - 1 7月 2022

指纹

探究 'Optimizing Constrained Guidance Policy with Minimum Overload Regularization' 的科研主题。它们共同构成独一无二的指纹。

引用此