Abstract

For batch processes with partial actuator failures and unknown system dynamics, an innovative two-dimensional (2D) model-free Q-learning algorithm is proposed to obtain the optimal controller's gains, achieving output feedback fault-tolerant control. First, a 2D linear model is constructed to describe batch processes with partial actuator failures. Then, the state increments in the batch direction and the output errors in the time direction are used as novel state variables to construct a multi-degree-of-freedom model. Second, a 2D Bellman equation is proposed through a connection between a 2D value and a 2D Q functions. Next, a 2D off-policy model-free Q-learning algorithm is highlighted, which incorporates target policies into a multi-degree-of-freedom model and focuses on using policy iteration to solve the fault-tolerant tracking control problem. The robustness analysis rigorously proves the stability of the closed-loop system. Lastly, the simulation results of the holding stage prove the feasibility and effectiveness of the presented algorithm.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.