TY - GEN
T1 - Deep Reinforcement Learning-Based Multi-constraint Guidance with Field-of-View Limitation
AU - Pu, Yuhui
AU - Bin, Yuru
AU - Wang, Hui
AU - Yang, Haorui
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - The advancement of deep reinforcement learning (DRL) algorithms has created new avenues to solve multi-constraint problems. Typically, multi-constraint advice entails maximizing the intended goal while taking into account several restrictions. The guidance module is able to optimize the guidance command in real time by acting as an intelligent entity that is capable of gathering information and input from its surroundings. In this paper, we develop a multi-stage guiding architecture using the prediction and correction concept as our foundation. This study decouples the guidance law to realize the impact angle constraints in the first stage. In the second step, we fitted a predictor to estimate the flight’s time-to-go. Additionally, by adjusting the bias term of the proportional navigation guidance law, we set a corrector to control the impact time and angle. Subsequently, we select a nonlinear optimization function in order to restrict the field-of-view. The study offers a technique for producing the multi-constraint guidance command using DRL.
AB - The advancement of deep reinforcement learning (DRL) algorithms has created new avenues to solve multi-constraint problems. Typically, multi-constraint advice entails maximizing the intended goal while taking into account several restrictions. The guidance module is able to optimize the guidance command in real time by acting as an intelligent entity that is capable of gathering information and input from its surroundings. In this paper, we develop a multi-stage guiding architecture using the prediction and correction concept as our foundation. This study decouples the guidance law to realize the impact angle constraints in the first stage. In the second step, we fitted a predictor to estimate the flight’s time-to-go. Additionally, by adjusting the bias term of the proportional navigation guidance law, we set a corrector to control the impact time and angle. Subsequently, we select a nonlinear optimization function in order to restrict the field-of-view. The study offers a technique for producing the multi-constraint guidance command using DRL.
KW - Deep reinforcement learning (DRL)
KW - Field-of-view limitation
KW - Multi-constraint guidance
KW - Prediction and correction
UR - https://www.scopus.com/pages/publications/105001015561
U2 - 10.1007/978-981-96-2216-0_6
DO - 10.1007/978-981-96-2216-0_6
M3 - Conference contribution
AN - SCOPUS:105001015561
SN - 9789819622153
T3 - Lecture Notes in Electrical Engineering
SP - 54
EP - 63
BT - Advances in Guidance, Navigation and Control - Proceedings of 2024 International Conference on Guidance, Navigation and Control Volume 5
A2 - Yan, Liang
A2 - Duan, Haibin
A2 - Deng, Yimin
PB - Springer Science and Business Media Deutschland GmbH
T2 - International Conference on Guidance, Navigation and Control, ICGNC 2024
Y2 - 9 August 2024 through 11 August 2024
ER -