Abstract

The present study establishes an approximate optimal critic learning algorithm, based on the single-network integral reinforcement learning (IRL) algorithm and intends to solve the optimal control problem for an unknown nonlinear system with saturating actuators. The value function is formulated through building generalized nonquadratic functions. In order to solve the Hamilton–Jacobi–Bellman (HJB) equation, a novel optimal scheme for the control approximation, based on the off-policy iteration is presented. Moreover, the single-neural network implementation procedure is introduced to complete the iteration algorithm. The synchronous IRL policy iteration is proposed to update the weight of the critic neural network. Finally, reasonable simulation results are provided for confirming the effectiveness of the proposed optimal approximation control technique in solving equations for a linear and oscillating systems.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call