TY - JOUR
T1 - Trajectory Tracking Control for Under-Actuated Hovercraft Using Differential Flatness and Reinforcement Learning-Based Active Disturbance Rejection Control
AU - Kong, Xiangyu
AU - Xia, Yuanqing
AU - Hu, Rui
AU - Lin, Min
AU - Sun, Zhongqi
AU - Dai, Li
N1 - Publisher Copyright:
© 2022, The Editorial Office of JSSC & Springer-Verlag GmbH Germany.
PY - 2022/4
Y1 - 2022/4
N2 - This paper proposes a scheme of trajectory tracking control for the hovercraft. Since the model of the hovercraft is under-actuated, nonlinear, and strongly coupled, it is a great challenge for the controller design. To solve this problem, the control scheme is divided into two parts. Firstly, we employ differential flatness method to find a set of flat outputs and consider part of the nonlinear terms as uncertainties. Consequently, we convert the under-actuated system into a full-actuated one. Secondly, a reinforcement learning-based active disturbance rejection controller (RL-ADRC) is designed. In this method, an extended state observer (ESO) is designed to estimate the uncertainties of the system, and an actorcritic-based reinforcement learning (RL) algorithm is used to approximate the optimal control strategy. Based on the output of the ESO, the RL-ADRC compensates for the total uncertainties in real-time, and simultaneously, generates the optimal control strategy by RL algorithm. Simulation results show that, compared with the traditional ADRC method, RL-ADRC does not need to manually tune the controller parameters, and the control strategy is more robust.
AB - This paper proposes a scheme of trajectory tracking control for the hovercraft. Since the model of the hovercraft is under-actuated, nonlinear, and strongly coupled, it is a great challenge for the controller design. To solve this problem, the control scheme is divided into two parts. Firstly, we employ differential flatness method to find a set of flat outputs and consider part of the nonlinear terms as uncertainties. Consequently, we convert the under-actuated system into a full-actuated one. Secondly, a reinforcement learning-based active disturbance rejection controller (RL-ADRC) is designed. In this method, an extended state observer (ESO) is designed to estimate the uncertainties of the system, and an actorcritic-based reinforcement learning (RL) algorithm is used to approximate the optimal control strategy. Based on the output of the ESO, the RL-ADRC compensates for the total uncertainties in real-time, and simultaneously, generates the optimal control strategy by RL algorithm. Simulation results show that, compared with the traditional ADRC method, RL-ADRC does not need to manually tune the controller parameters, and the control strategy is more robust.
KW - Active disturbance rejection control
KW - differential flatness
KW - reinforcement learning
KW - trajectory tracking control
KW - under-actuated system
UR - http://www.scopus.com/inward/record.url?scp=85128318648&partnerID=8YFLogxK
U2 - 10.1007/s11424-022-2037-0
DO - 10.1007/s11424-022-2037-0
M3 - Article
AN - SCOPUS:85128318648
SN - 1009-6124
VL - 35
SP - 502
EP - 521
JO - Journal of Systems Science and Complexity
JF - Journal of Systems Science and Complexity
IS - 2
ER -