Abstract

To eliminate the requirement of the precise model for model-based control methods, an improved modelfree H∞ control method is designed for batch processes with unknown dynamics and disturbances. Firstly, both the zero-sum game value function and the Q-function are presented as two-dimensional (2D) forms, and their relation is analyzed to obtain the model-free Bellman equation. Secondly, an on policy 2D game Q-learning method is proposed for learning the optimal gains of the designed H∞ controller. On this basis, the behavior control policy and the behavior disturbance policy are individually applied by developing an off-policy 2D game Q-learning method. Subsequently, the strict proof about the convergence and the unbiasedness of the off-policy approach are given. Finally, the simulation results of injection velocity manifest the validity and effectivity of the proposed algorithm.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call