TY - JOUR
T1 - Adaptive Fixed-Time Optimal Formation Control for Uncertain Nonlinear Multiagent Systems Using Reinforcement Learning
AU - Wang, Ping
AU - Yu, Chengpu
AU - Lv, Maolong
AU - Cao, Jinde
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2024/3/1
Y1 - 2024/3/1
N2 - This article explores the application of reinforcement learning (RL) strategy to achieve an adaptive fixed-time (FxT) optimized formation control of uncertain nonlinear multiagent systems. The primary obstacle in this process is the difficulty in attaining FxT stability under the actor-critic setting due to intermediate estimation errors and generic system uncertainties. To overcome these challenges, the RL control algorithm is implemented using an identifier-actor-critic structure, where the identifier is utilized to address the system uncertainty involving unknown nonlinear dynamics and external disturbances. Furthermore, a novel quadratic function is introduced to establish the boundedness of the estimation error of the actor-critic learning law, which plays a pivotal role in the FxT stability analysis. Finally, a unified FxT optimized formation control strategy is developed, which guarantees the realization of the predetermined formation at a fixed time while optimizing the given performance measure. The effectiveness of the proposed control algorithm is verified through simulation of a team of marine surface vessels.
AB - This article explores the application of reinforcement learning (RL) strategy to achieve an adaptive fixed-time (FxT) optimized formation control of uncertain nonlinear multiagent systems. The primary obstacle in this process is the difficulty in attaining FxT stability under the actor-critic setting due to intermediate estimation errors and generic system uncertainties. To overcome these challenges, the RL control algorithm is implemented using an identifier-actor-critic structure, where the identifier is utilized to address the system uncertainty involving unknown nonlinear dynamics and external disturbances. Furthermore, a novel quadratic function is introduced to establish the boundedness of the estimation error of the actor-critic learning law, which plays a pivotal role in the FxT stability analysis. Finally, a unified FxT optimized formation control strategy is developed, which guarantees the realization of the predetermined formation at a fixed time while optimizing the given performance measure. The effectiveness of the proposed control algorithm is verified through simulation of a team of marine surface vessels.
KW - Optimal control
KW - fixed-time stability
KW - formation control
KW - marine surface vessels
KW - reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85177058146&partnerID=8YFLogxK
U2 - 10.1109/TNSE.2023.3330266
DO - 10.1109/TNSE.2023.3330266
M3 - Article
AN - SCOPUS:85177058146
SN - 2327-4697
VL - 11
SP - 1729
EP - 1743
JO - IEEE Transactions on Network Science and Engineering
JF - IEEE Transactions on Network Science and Engineering
IS - 2
ER -