Abstract
Aiming at addressing the problem of manoeuvring decision-making in UAV air combat, this study establishes a one-to-one air combat model, defines missile attack areas, and uses the non-deterministic policy Soft-Actor-Critic (SAC) algorithm in deep reinforcement learning to construct a decision model to realize the manoeuvring process. At the same time, the complexity of the proposed algorithm is calculated, and the stability of the closed-loop system of air combat decision-making controlled by neural network is analysed by the Lyapunov function. This study defines the UAV air combat process as a gaming process and proposes a Parallel Self-Play training SAC algorithm (PSP-SAC) to improve the generalisation performance of UAV control decisions. Simulation results have shown that the proposed algorithm can realize sample sharing and policy sharing in multiple combat environments and can significantly improve the generalisation ability of the model compared to independent training.
Original language | English |
---|---|
Pages (from-to) | 64-81 |
Number of pages | 18 |
Journal | CAAI Transactions on Intelligence Technology |
Volume | 8 |
Issue number | 1 |
DOIs | |
Publication status | Published - Mar 2023 |
Keywords
- SAC algorithm
- UAV
- air combat decision
- deep reinforcement learning
- parallel self-play