Abstract
AbstractThe optimal control issue of discrete‐time nonlinear unknown systems with time‐delay control input is the focus of this work. In order to reduce communication costs, a reinforcement learning‐based event‐triggered controller is proposed. By applying the proposed control method, closed‐loop system's asymptotic stability is demonstrated, and a maximum upper bound for the infinite‐horizon performance index can be calculated beforehand. The event‐triggered condition requires the next time state information. In an effort to forecast the next state and achieve optimal control, three neural networks (NNs) are introduced and used to approximate system state, value function, and optimal control. Additionally, a M NN is utilized to cope with the time‐delay term of control input. Moreover, taking the estimation errors of NNs into account, the uniformly ultimately boundedness of state and NNs weight estimation errors can be guaranteed. Ultimately, the validity of proposed approach is illustrated by simulations.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: International Journal of Robust and Nonlinear Control
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.