Deep reinforcement learning and its application in autonomous fitting optimization for attack areas of UCAVs

Li Yue, Qiu Xiaohui*, Liu Xiaodong, Xia Qunli

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

31 Citations (Scopus)

Abstract

The ever-changing battlefield environment requires the use of robust and adaptive technologies integrated into a reliable platform. Unmanned combat aerial vehicles (UCAVs) aim to integrate such advanced technologies while increasing the tactical capabilities of combat aircraft. As a research object, common UCAV uses the neural network fitting strategy to obtain values of attack areas. However, this simple strategy cannot cope with complex environmental changes and autonomously optimize decision-making problems. To solve the problem, this paper proposes a new deep deterministic policy gradient (DDPG) strategy based on deep reinforcement learning for the attack area fitting of UCAVs in the future battlefield. Simulation results show that the autonomy and environmental adaptability of UCAVs in the future battlefield will be improved based on the new DDPG algorithm and the training process converges quickly. We can obtain the optimal values of attack areas in real time during the whole flight with the well-trained deep network.

Original languageEnglish
Pages (from-to)734-742
Number of pages9
JournalJournal of Systems Engineering and Electronics
Volume31
Issue number4
DOIs
Publication statusPublished - Aug 2020

Keywords

  • attack area
  • deep deterministic policy gradient (DDPG)
  • neural network
  • unmanned combat aerial vehicle (UCAV)

Fingerprint

Dive into the research topics of 'Deep reinforcement learning and its application in autonomous fitting optimization for attack areas of UCAVs'. Together they form a unique fingerprint.

Cite this