TY - JOUR
T1 - Multi-UAV Cooperative Search Based on Reinforcement Learning With a Digital Twin Driven Training Framework
AU - Shen, Gaoqing
AU - Lei, Lei
AU - Zhang, Xinting
AU - Li, Zhilin
AU - Cai, Shengsuo
AU - Zhang, Lijuan
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023/7/1
Y1 - 2023/7/1
N2 - This paper considers the cooperative search for stationary targets by multiple unmanned aerial vehicles (UAVs) with limited sensing range and communication ability in a dynamic threatening environment. The main purpose is to use multiple UAVs to find more unknown targets as soon as possible, increase the coverage rate of the mission area, and more importantly, guide UAVs away from threats. However, traditional search methods are mostly unscalable and perform poorly in dynamic environments. A new multi-agent deep reinforcement learning (MADRL) method, DNQMIX, is proposed in this study to solve the multi-UAV cooperative target search (MCTS) problem. The reward function is also newly designed for the MCTS problem to guide UAVs to explore and exploit the environment information more efficiently. Moreover, this paper proposes a digital twin (DT) driven training framework 'centralized training, decentralized execution, and continuous evolution' (CTDECE). It can facilitate the continuous evolution of MADRL models and solve the tradeoff between training speed and environment fidelity when MADRL is applied to real-world multi-UAV systems. Simulation results show that DNQMIX outperforms state-of-art methods in terms of search rate and coverage rate.
AB - This paper considers the cooperative search for stationary targets by multiple unmanned aerial vehicles (UAVs) with limited sensing range and communication ability in a dynamic threatening environment. The main purpose is to use multiple UAVs to find more unknown targets as soon as possible, increase the coverage rate of the mission area, and more importantly, guide UAVs away from threats. However, traditional search methods are mostly unscalable and perform poorly in dynamic environments. A new multi-agent deep reinforcement learning (MADRL) method, DNQMIX, is proposed in this study to solve the multi-UAV cooperative target search (MCTS) problem. The reward function is also newly designed for the MCTS problem to guide UAVs to explore and exploit the environment information more efficiently. Moreover, this paper proposes a digital twin (DT) driven training framework 'centralized training, decentralized execution, and continuous evolution' (CTDECE). It can facilitate the continuous evolution of MADRL models and solve the tradeoff between training speed and environment fidelity when MADRL is applied to real-world multi-UAV systems. Simulation results show that DNQMIX outperforms state-of-art methods in terms of search rate and coverage rate.
KW - Cooperative target search
KW - digital twin
KW - multi-agent deep reinforcement learning
KW - unmanned aerial vehicles
UR - http://www.scopus.com/inward/record.url?scp=85149414194&partnerID=8YFLogxK
U2 - 10.1109/TVT.2023.3245120
DO - 10.1109/TVT.2023.3245120
M3 - Article
AN - SCOPUS:85149414194
SN - 0018-9545
VL - 72
SP - 8354
EP - 8368
JO - IEEE Transactions on Vehicular Technology
JF - IEEE Transactions on Vehicular Technology
IS - 7
ER -