Abstract

This article investigates an optimized containment control problem for multiagent systems (MASs), where all followers are subject to deferred full-state constraints. A universal nonlinear transformation is proposed for simultaneously handling the cases with and without constraints. Particularly, for the constrained case, initial values of states are flexibly managed to the midpoint between upper and lower boundaries by utilizing a state-shifting function, thus eliminating the initial restriction conditions. By deferred constraints, the state is forced to fall back into the restrictive boundaries within a preassigned time. A neural network (NN)-based reinforcement learning (RL) algorithm is executed under the identifier-critic-actor architecture, where the Hamilton-Jacobi-Bellman (HJB) equation is built in every subsystem to optimize control performance. For actor and critic NNs, updating laws are simplified, since the gradient descent method is performed based on a simple positive function rather than square of Bellman residual error. In view of the Lyapunov stability theorem and graph theory, it is proved that all signals are bounded and the outputs of followers can eventually enter into the convex hull constituted by leaders. Finally, simulations confirm the validity of the proposed approach.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.