Abstract
The ever-changing battlefield environment requires the use of robust and adaptive technologies integrated into a reliable platform. Unmanned combat aerial vehicles (UCAVs) aim to integrate such advanced technologies while increasing the tactical capabilities of combat aircraft. As a research object, common UCAV uses the neural network fitting strategy to obtain values of attack areas. However, this simple strategy cannot cope with complex environmental changes and autonomously optimize decision-making problems. To solve the problem, this paper proposes a new deep deterministic policy gradient (DDPG) strategy based on deep reinforcement learning for the attack area fitting of UCAVs in the future battlefield. Simulation results show that the autonomy and environmental adaptability of UCAVs in the future battlefield will be improved based on the new DDPG algorithm and the training process converges quickly. We can obtain the optimal values of attack areas in real time during the whole flight with the well-trained deep network.
Original language | English |
---|---|
Pages (from-to) | 734-742 |
Number of pages | 9 |
Journal | Journal of Systems Engineering and Electronics |
Volume | 31 |
Issue number | 4 |
DOIs | |
Publication status | Published - Aug 2020 |
Keywords
- attack area
- deep deterministic policy gradient (DDPG)
- neural network
- unmanned combat aerial vehicle (UCAV)