TY - JOUR
T1 - A Smooth Path Planning Learning Strategy Design for an Air-Ground Vehicle Considering Mode Switching
AU - Zhao, Jing
AU - Yang, Chao
AU - Wang, Weida
AU - Li, Ying
AU - Xiang, Changle
N1 - Publisher Copyright:
© 2021 ICAE.
PY - 2021
Y1 - 2021
N2 - With the ability of vertical take-off and landing, the task path of an air-ground vehicle will be significantly shortened. Accordingly, the energy consumption will be greatly reduced. Through reasonable planning of the path, such vehicle can meet the high-efficiency needs of unmanned tasks and alleviate the global energy shortage problem. To design an optimal feasible path, this paper proposes a smooth path planning learning strategy considering mode switching. A new reward function of the Q-learning algorithm is presented, considering the influence of flight obstacle crossing parameters. To avoid the redundant flight distance and energy consumption caused by frequent high flights, the flight height correction is made in the update rule. Besides that, a path smoothing modification, called double yaw correction, reduces turning points and improves the path smoothness. It further reduces the energy consumption caused by the tortuous path. This modification also points out the direction of iterative learning and accelerates the algorithm convergence speed. Finally, the proposed strategy is verified on a 40m*40m map with 0-10m obstacle height. Results show that, the proposed strategy is effective to shorten 4.08m distance and plays the role of smoothing the path. Its convergence speed is faster than the traditional algorithm.
AB - With the ability of vertical take-off and landing, the task path of an air-ground vehicle will be significantly shortened. Accordingly, the energy consumption will be greatly reduced. Through reasonable planning of the path, such vehicle can meet the high-efficiency needs of unmanned tasks and alleviate the global energy shortage problem. To design an optimal feasible path, this paper proposes a smooth path planning learning strategy considering mode switching. A new reward function of the Q-learning algorithm is presented, considering the influence of flight obstacle crossing parameters. To avoid the redundant flight distance and energy consumption caused by frequent high flights, the flight height correction is made in the update rule. Besides that, a path smoothing modification, called double yaw correction, reduces turning points and improves the path smoothness. It further reduces the energy consumption caused by the tortuous path. This modification also points out the direction of iterative learning and accelerates the algorithm convergence speed. Finally, the proposed strategy is verified on a 40m*40m map with 0-10m obstacle height. Results show that, the proposed strategy is effective to shorten 4.08m distance and plays the role of smoothing the path. Its convergence speed is faster than the traditional algorithm.
KW - Q reinforcement learning
KW - air-ground vehicle
KW - mode switching
KW - path planning
KW - path smoothness
UR - http://www.scopus.com/inward/record.url?scp=85190418359&partnerID=8YFLogxK
U2 - 10.46855/energy-proceedings-9228
DO - 10.46855/energy-proceedings-9228
M3 - Conference article
AN - SCOPUS:85190418359
SN - 2004-2965
VL - 18
JO - Energy Proceedings
JF - Energy Proceedings
T2 - 13th International Conference on Applied Energy, ICAE 2021
Y2 - 29 November 2021 through 2 December 2021
ER -