TY - CHAP
T1 - Robust Formation Tracking Control for Multi-Agent Systems Using Reinforcement Learning Methods
AU - Dong, Xiwang
AU - Shi, Yu
AU - Hua, Yongzhao
AU - Yu, Jianglong
AU - Ren, Zhang
N1 - Publisher Copyright:
© 2026 Elsevier Inc. All rights are reserved, including those for text and data mining, AI training, and similar technologies.
PY - 2025/1/1
Y1 - 2025/1/1
N2 - This chapter investigates a fully data-driven method to solve the robust output formation tracking control problem for the multi-agent system (MAS) under actuator faults. The outputs of the followers are controlled to track those of multiple leaders with respect to a convex point while achieving an expected time-varying formation. To obviate the requirement of various system prior knowledge in typical MAS control, a hierarchical framework is developed with three learning and control stages using the online measured data. First, a distributed adaptive observer is designed to coordinate the state convex of multiple leaders while estimating unknown dynamics. The adaptive mechanism relaxes the demand for global topology. Second, by collecting and reusing the online system data, an off-policy reinforcement learning (RL) method is proposed in a continuous form to acquire nominal feedback gains from partial observations of the followers. Essential system models are learned along with the RL process, while solutions to the output regulation equations are implicitly obtained. Third, a comprehensive robust controller is further presented based on the previous learning results. To address the actuator faults with efficiency loss and bias, the adaptive neural networks and robust compensations are utilized in a model-free manner. The output formation tracking is achieved under a derived feasibility condition while stabilities of the learning and control methods are analyzed. Finally, simulation results demonstrate the validity of this fully data-driven control framework.
AB - This chapter investigates a fully data-driven method to solve the robust output formation tracking control problem for the multi-agent system (MAS) under actuator faults. The outputs of the followers are controlled to track those of multiple leaders with respect to a convex point while achieving an expected time-varying formation. To obviate the requirement of various system prior knowledge in typical MAS control, a hierarchical framework is developed with three learning and control stages using the online measured data. First, a distributed adaptive observer is designed to coordinate the state convex of multiple leaders while estimating unknown dynamics. The adaptive mechanism relaxes the demand for global topology. Second, by collecting and reusing the online system data, an off-policy reinforcement learning (RL) method is proposed in a continuous form to acquire nominal feedback gains from partial observations of the followers. Essential system models are learned along with the RL process, while solutions to the output regulation equations are implicitly obtained. Third, a comprehensive robust controller is further presented based on the previous learning results. To address the actuator faults with efficiency loss and bias, the adaptive neural networks and robust compensations are utilized in a model-free manner. The output formation tracking is achieved under a derived feasibility condition while stabilities of the learning and control methods are analyzed. Finally, simulation results demonstrate the validity of this fully data-driven control framework.
KW - Adaptive neural networks.
KW - Data-driven control
KW - Fault-tolerant
KW - Multi-agent system
KW - Output formation tracking
KW - Partial observations
KW - Reinforcement learning
UR - https://www.scopus.com/pages/publications/105028837797
U2 - 10.1016/B978-0-443-14081-5.00128-8
DO - 10.1016/B978-0-443-14081-5.00128-8
M3 - 章节
AN - SCOPUS:105028837797
SN - 9780443140808
SP - V3:573-V3:592
BT - Encyclopedia of Systems and Control Engineering
PB - Elsevier
ER -