TY - JOUR
T1 - Learning Hybrid Policies for MPC with Application to Drone Flight in Unknown Dynamic Environments
AU - Feng, Zhaohan
AU - Chen, Jie
AU - Xiao, Wei
AU - Sun, Jian
AU - Xin, Bin
AU - Wang, Gang
N1 - Publisher Copyright:
#c World Scientific Publishing Company.
PY - 2024/3/1
Y1 - 2024/3/1
N2 - In recent years, drones have found increased applications in a wide array of real-world tasks. Model predictive control (MPC) has emerged as a practical method for drone flight control, owing to its robustness against modeling errors/uncertainties and external disturbances. However, MPC’s sensitivity to manually tuned parameters can lead to rapid performance degradation when faced with unknown environmental dynamics. This paper addresses the challenge of controlling a drone as it traverses a swinging gate characterized by unknown dynamics. This paper introduces a parameterized MPC approach named hyMPC that leverages high-level decision variables to adapt to uncertain environmental conditions. To derive these decision variables, a novel policy search framework aimed at training a high-level Gaussian policy is presented. Subsequently, we harness the power of neural network policies, trained on data gathered through the repeated execution of the Gaussian policy, to provide real-time decision variables. The effectiveness of hyMPC is validated through numerical simulations, achieving a 100% success rate in 20 drone flight tests traversing a swinging gate, demonstrating its capability to achieve safe and precise flight with limited prior knowledge of environmental dynamics.
AB - In recent years, drones have found increased applications in a wide array of real-world tasks. Model predictive control (MPC) has emerged as a practical method for drone flight control, owing to its robustness against modeling errors/uncertainties and external disturbances. However, MPC’s sensitivity to manually tuned parameters can lead to rapid performance degradation when faced with unknown environmental dynamics. This paper addresses the challenge of controlling a drone as it traverses a swinging gate characterized by unknown dynamics. This paper introduces a parameterized MPC approach named hyMPC that leverages high-level decision variables to adapt to uncertain environmental conditions. To derive these decision variables, a novel policy search framework aimed at training a high-level Gaussian policy is presented. Subsequently, we harness the power of neural network policies, trained on data gathered through the repeated execution of the Gaussian policy, to provide real-time decision variables. The effectiveness of hyMPC is validated through numerical simulations, achieving a 100% success rate in 20 drone flight tests traversing a swinging gate, demonstrating its capability to achieve safe and precise flight with limited prior knowledge of environmental dynamics.
KW - Model predictive control
KW - reinforcement learning
KW - trajectory planning
KW - unmanned aerial vehicle
UR - http://www.scopus.com/inward/record.url?scp=85187506873&partnerID=8YFLogxK
U2 - 10.1142/S2301385024410206
DO - 10.1142/S2301385024410206
M3 - Article
AN - SCOPUS:85187506873
SN - 2301-3850
VL - 12
SP - 429
EP - 441
JO - Unmanned Systems
JF - Unmanned Systems
IS - 2
ER -