TY - JOUR
T1 - Optimizing Constrained Guidance Policy with Minimum Overload Regularization
AU - Luo, Weilin
AU - Chen, Lei
AU - Liu, Kexin
AU - Gu, Haibo
AU - Lu, Jinhu
N1 - Publisher Copyright:
© 2004-2012 IEEE.
PY - 2022/7/1
Y1 - 2022/7/1
N2 - Using reinforcement learning (RL) algorithm to optimize guidance law can address non-idealities in complex environment. However, the optimization is difficult due to huge state-Action space, unstable training, and high requirements on expertise. In this paper, the constrained guidance policy of a neural guidance system is optimized using improved RL algorithm, which is motivated by the idea of traditional model-based guidance method. A novel optimization objective with minimum overload regularization is developed to restrain the guidance policy directly from generating redundant missile maneuver. Moreover, a bi-level curriculum learning is designed to facilitate the policy optimization. Experiment results show that the proposed minimum overload regularization can reduce the vertical overloads of missile significantly, and the bi-level curriculum learning can further accelerate the optimization of guidance policy.
AB - Using reinforcement learning (RL) algorithm to optimize guidance law can address non-idealities in complex environment. However, the optimization is difficult due to huge state-Action space, unstable training, and high requirements on expertise. In this paper, the constrained guidance policy of a neural guidance system is optimized using improved RL algorithm, which is motivated by the idea of traditional model-based guidance method. A novel optimization objective with minimum overload regularization is developed to restrain the guidance policy directly from generating redundant missile maneuver. Moreover, a bi-level curriculum learning is designed to facilitate the policy optimization. Experiment results show that the proposed minimum overload regularization can reduce the vertical overloads of missile significantly, and the bi-level curriculum learning can further accelerate the optimization of guidance policy.
KW - Missile guidance
KW - curriculum learning
KW - minimum overload regularization
KW - reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85133563279&partnerID=8YFLogxK
U2 - 10.1109/TCSI.2022.3163463
DO - 10.1109/TCSI.2022.3163463
M3 - Article
AN - SCOPUS:85133563279
SN - 1549-8328
VL - 69
SP - 2994
EP - 3005
JO - IEEE Transactions on Circuits and Systems I: Regular Papers
JF - IEEE Transactions on Circuits and Systems I: Regular Papers
IS - 7
ER -