TY - GEN
T1 - An Improved DDPG Algorithm for UAV Navigation in Large-Scale Complex Environments
AU - Peng, Jiabin
AU - Lv, Bo
AU - Zhang, Lijuan
AU - Lei, Lei
AU - Song, Xiaoqin
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Autonomous navigation is one of the most critical elementary skills to control UAV navigate around the environment without colliding with obstacles. As UAVs are prevalently applied in many domains, such as goods delivery, search and rescue and intelligent transportation etc, safety and efficiency are two basic requirements while executing tasks. However, autonomous navigation is not a trivial task because it is challenging for UAV to observe, orientate, decide and take actions simultaneously. Specially, in large-scale complex environments, the constrained narrow passages, dense obstacles and dynamic objects pose increasing difficulty. In this article, UAV autonomous navigation is modeled as a Markov decision process and an Improved Deep Deterministic Policy Gradient (ImDDPG) algorithm is proposed to make safe and efficient navigation decisions. In ImDDPG, the actor network is decomposed into two sub-networks to learn a more stable action policy. Next, some reward reshaping functions are developed to solve the sparse reward problem. Then, the convergence speed is further accelerated with a dynamic decay strategy. The targeting neural network is well designed and trained. Simulation experiments are conducted to prove that the proposed ImDDPG algorithm can achieve a success rate more than 95% in large-scale complex environments. Moreover, numerous simulation results are also presented to demonstrate that ImDDPG has better generalization ability than the benchmark DDPG and TD3 algorithms in environments with denser obstacles and dynamic objects.
AB - Autonomous navigation is one of the most critical elementary skills to control UAV navigate around the environment without colliding with obstacles. As UAVs are prevalently applied in many domains, such as goods delivery, search and rescue and intelligent transportation etc, safety and efficiency are two basic requirements while executing tasks. However, autonomous navigation is not a trivial task because it is challenging for UAV to observe, orientate, decide and take actions simultaneously. Specially, in large-scale complex environments, the constrained narrow passages, dense obstacles and dynamic objects pose increasing difficulty. In this article, UAV autonomous navigation is modeled as a Markov decision process and an Improved Deep Deterministic Policy Gradient (ImDDPG) algorithm is proposed to make safe and efficient navigation decisions. In ImDDPG, the actor network is decomposed into two sub-networks to learn a more stable action policy. Next, some reward reshaping functions are developed to solve the sparse reward problem. Then, the convergence speed is further accelerated with a dynamic decay strategy. The targeting neural network is well designed and trained. Simulation experiments are conducted to prove that the proposed ImDDPG algorithm can achieve a success rate more than 95% in large-scale complex environments. Moreover, numerous simulation results are also presented to demonstrate that ImDDPG has better generalization ability than the benchmark DDPG and TD3 algorithms in environments with denser obstacles and dynamic objects.
UR - http://www.scopus.com/inward/record.url?scp=85160562246&partnerID=8YFLogxK
U2 - 10.1109/AERO55745.2023.10115997
DO - 10.1109/AERO55745.2023.10115997
M3 - Conference contribution
AN - SCOPUS:85160562246
T3 - IEEE Aerospace Conference Proceedings
BT - 2023 IEEE Aerospace Conference, AERO 2023
PB - IEEE Computer Society
T2 - 2023 IEEE Aerospace Conference, AERO 2023
Y2 - 4 March 2023 through 11 March 2023
ER -