TY - JOUR
T1 - A State-Decomposition DDPG Algorithm for UAV Autonomous Navigation in 3-D Complex Environments
AU - Zhang, Lijuan
AU - Peng, Jiabin
AU - Yi, Weiguo
AU - Lin, Hang
AU - Lei, Lei
AU - Song, Xiaoqin
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024/3/15
Y1 - 2024/3/15
N2 - Over the past decade, unmanned aerial vehicles (UAVs) have been widely applied in many areas, such as goods delivery, disaster monitoring, search and rescue etc. In most of these applications, autonomous navigation is one of the key techniques that enable UAV to perform various tasks. However, UAV autonomous navigation in complex environments presents significant challenges due to the difficulty in simultaneously observing, orientation, decision and action. In this work, an efficient state-decomposition deep deterministic policy gradient algorithm is proposed for UAV autonomous navigation (SDDPG-NAV) in 3-D complex environments. In SDDPG-NAV, a novel state-decomposition method that uses two subnetworks for the perception-related and target-related states separately is developed to establish more appropriate actor networks. We also designed some objective-oriented reward functions to solve the sparse reward problem, including approaching the target, and avoiding obstacles and step award functions. Moreover, some training strategies are introduced to maintain the balance between exploration and exploitation, and the network is well trained with numerous experiments. The proposed SDDPG-NAV algorithm is capable of adapting to surrounding environments with generalized training experiences and effectively improves UAV's navigation performance in 3-D complex environments. Comparing with the benchmark DDPG and TD3 algorithms, SDDPG-NAV exhibits better performance in terms of convergence rate, navigation performance, and generalization capability.
AB - Over the past decade, unmanned aerial vehicles (UAVs) have been widely applied in many areas, such as goods delivery, disaster monitoring, search and rescue etc. In most of these applications, autonomous navigation is one of the key techniques that enable UAV to perform various tasks. However, UAV autonomous navigation in complex environments presents significant challenges due to the difficulty in simultaneously observing, orientation, decision and action. In this work, an efficient state-decomposition deep deterministic policy gradient algorithm is proposed for UAV autonomous navigation (SDDPG-NAV) in 3-D complex environments. In SDDPG-NAV, a novel state-decomposition method that uses two subnetworks for the perception-related and target-related states separately is developed to establish more appropriate actor networks. We also designed some objective-oriented reward functions to solve the sparse reward problem, including approaching the target, and avoiding obstacles and step award functions. Moreover, some training strategies are introduced to maintain the balance between exploration and exploitation, and the network is well trained with numerous experiments. The proposed SDDPG-NAV algorithm is capable of adapting to surrounding environments with generalized training experiences and effectively improves UAV's navigation performance in 3-D complex environments. Comparing with the benchmark DDPG and TD3 algorithms, SDDPG-NAV exhibits better performance in terms of convergence rate, navigation performance, and generalization capability.
KW - Autonomous navigation
KW - decision making
KW - deep reinforcement learning (DRL)
KW - path planning
KW - unmanned aerial vehicle (UAV) autonomy
UR - http://www.scopus.com/inward/record.url?scp=85181560082&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2023.3327753
DO - 10.1109/JIOT.2023.3327753
M3 - Article
AN - SCOPUS:85181560082
SN - 2327-4662
VL - 11
SP - 10778
EP - 10790
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 6
ER -