TY - JOUR
T1 - A dynamic control decision approach for fixed-wing aircraft games via hybrid action reinforcement learning
AU - Zhuang, Xing
AU - Li, Dongguang
AU - Li, Hanyu
AU - Wang, Yue
AU - Zhu, Jihong
N1 - Publisher Copyright:
© Science China Press 2025.
PY - 2025/3
Y1 - 2025/3
N2 - Autonomous decision-making is crucial for aircraft to achieve quick victories in diverse scenarios. Based on a 6-degree-of-freedom aircraft model, this paper proposes a decoupled guidance and control theory for autonomous aircraft maneuvering, distinguishing between close and long-range engagements. We introduce a method for heading attitude control to enhance stability during close-range interactions and a speed-based adaptive grid model for precise waypoint control in mid-to-long-range engagements. The paper transforms dynamic aircraft interactions into a Markov decision process and presents a hybrid discrete and continuous action reinforcement learning approach. This unified learning framework offers enhanced generalization and learning speed for dynamic aircraft adversarial processes. Experimental results indicate that in a symmetric environment, our approach rapidly achieves Nash equilibrium, securing over a 10% advantage. In unmanned aerial aircraft game control with higher maneuverability, the probability of gaining a situational advantage increases by more than 40%. Compared to similar methods, our approach demonstrates superior effectiveness in decision optimization and adversarial success probability. Furthermore, we validate the algorithm’s robustness and adaptability in an asymmetric environment, showcasing its promising application potential in collaborative control of aircraft clusters.
AB - Autonomous decision-making is crucial for aircraft to achieve quick victories in diverse scenarios. Based on a 6-degree-of-freedom aircraft model, this paper proposes a decoupled guidance and control theory for autonomous aircraft maneuvering, distinguishing between close and long-range engagements. We introduce a method for heading attitude control to enhance stability during close-range interactions and a speed-based adaptive grid model for precise waypoint control in mid-to-long-range engagements. The paper transforms dynamic aircraft interactions into a Markov decision process and presents a hybrid discrete and continuous action reinforcement learning approach. This unified learning framework offers enhanced generalization and learning speed for dynamic aircraft adversarial processes. Experimental results indicate that in a symmetric environment, our approach rapidly achieves Nash equilibrium, securing over a 10% advantage. In unmanned aerial aircraft game control with higher maneuverability, the probability of gaining a situational advantage increases by more than 40%. Compared to similar methods, our approach demonstrates superior effectiveness in decision optimization and adversarial success probability. Furthermore, we validate the algorithm’s robustness and adaptability in an asymmetric environment, showcasing its promising application potential in collaborative control of aircraft clusters.
KW - dynamic control
KW - intelligent air combat
KW - reinforcement learning
KW - unmanned aerial vehicle game
UR - http://www.scopus.com/inward/record.url?scp=85218355211&partnerID=8YFLogxK
U2 - 10.1007/s11432-023-4217-8
DO - 10.1007/s11432-023-4217-8
M3 - Article
AN - SCOPUS:85218355211
SN - 1674-733X
VL - 68
JO - Science China Information Sciences
JF - Science China Information Sciences
IS - 3
M1 - 132201
ER -