TY - JOUR
T1 - 非对称机动能力多无人机智能协同攻防对抗
AU - Chen, Can
AU - Mo, Li
AU - Zheng, Duo
AU - Cheng, Ziheng
AU - Lin, Defu
N1 - Publisher Copyright:
© 2020, Beihang University Aerospace Knowledge Press. All right reserved.
PY - 2020/12/25
Y1 - 2020/12/25
N2 - The attack-defense game is an important combat scenario of future military Unmanned Aerial Vehicles (UAVs). This paper studies an attack-defense game between groups of UAVs with different maneuverability, establishing a multi-UAV cooperative attack and defense evolution model. Based on the multi-agent reinforcement learning theory, the autonomous decision-making method of multi-UAV cooperative attack-defense game is studied, and a centralized critic and distributed actor algorithm structure is proposed based on the actor-critic algorithm, guaranteeing the convergence of the algorithm and improving the efficiency of decision-making. The critic module of UAVs uses the global information to evaluate the decision-making quality during training, while the actor module only needs to rely on the local perception information to make autonomous decisions during execution, hence improving the effectiveness of the multi-UAV attack-defense game. The simulation results show that the proposed multi-UAV reinforcement learning method has a strong self-evolution property, endowing the UAV certain intelligence, that is, the stable autonomous learning ability. Through continuous training, the UAVs can autonomously learn cooperative attack or defense policies to improve the effectiveness of decision-making.
AB - The attack-defense game is an important combat scenario of future military Unmanned Aerial Vehicles (UAVs). This paper studies an attack-defense game between groups of UAVs with different maneuverability, establishing a multi-UAV cooperative attack and defense evolution model. Based on the multi-agent reinforcement learning theory, the autonomous decision-making method of multi-UAV cooperative attack-defense game is studied, and a centralized critic and distributed actor algorithm structure is proposed based on the actor-critic algorithm, guaranteeing the convergence of the algorithm and improving the efficiency of decision-making. The critic module of UAVs uses the global information to evaluate the decision-making quality during training, while the actor module only needs to rely on the local perception information to make autonomous decisions during execution, hence improving the effectiveness of the multi-UAV attack-defense game. The simulation results show that the proposed multi-UAV reinforcement learning method has a strong self-evolution property, endowing the UAV certain intelligence, that is, the stable autonomous learning ability. Through continuous training, the UAVs can autonomously learn cooperative attack or defense policies to improve the effectiveness of decision-making.
KW - Attack-defense games
KW - Centralized critic
KW - Distributed actors
KW - Multi-UAV coordination
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85098990144&partnerID=8YFLogxK
U2 - 10.7527/S1000-6893.2020.24152
DO - 10.7527/S1000-6893.2020.24152
M3 - 文章
AN - SCOPUS:85098990144
SN - 1000-6893
VL - 41
JO - Hangkong Xuebao/Acta Aeronautica et Astronautica Sinica
JF - Hangkong Xuebao/Acta Aeronautica et Astronautica Sinica
IS - 12
M1 - 324152
ER -