Abstract

This paper presents a novel off-policy game Q-learning algorithm to solve $H_\infty $ control problem for discrete-time linear multi-player systems with completely unknown system dynamics. The primary contribution of this paper lies in that the Q-learning strategy employed in the proposed algorithm is implemented in an off-policy policy iteration approach other than on-policy learning, since the off-policy learning has some well-known advantages over the on-policy learning. All of players struggle together to minimize their common performance index meanwhile defeating the disturbance that tries to maximize the specific performance index, and finally they reach the Nash equilibrium of game resulting in satisfying disturbance attenuation condition. For finding the solution of the Nash equilibrium, $H_\infty $ control problem is first transformed into an optimal control problem. Then an off-policy Q-learning algorithm is put forward in the typical adaptive dynamic programming (ADP) and game architecture, such that control policies of all players can be learned using only measured data. More importantly, the rigorous proof of no bias of solution to the Nash equilibrium by using the proposed off-policy game Q-learning algorithm is presented. Comparative simulation results are provided to verify the effectiveness and demonstrate the advantages of the proposed method.

Highlights

  • The H∞ control is a robust control method which is aimed at designing the controllers to attenuate the negative effects in performance of dynamical systems caused by external disturbances guarantee the stability of systems if no disturbance exists [1]–[3]

  • By reviewing the existing results on H∞ control for dynamical systems, it is not difficult to find that most of researchers are concerned about model-based H∞ controller design using the variety of methods, such as linear matrix inequality (LMI) [10]–[12], zero-sum game [13]–[17] and pole assignment [18]–[20], etc

  • In view of the advantages of off-policy learning over on-policy learning shown in our previous result [37] wherein the off -policy Q-learning method was proposed for multi-player systems without the consideration of disturbance, developing an off-policy game Q-learning algorithm to solve H∞ control problem for discrete-time linear multi-player systems using only measured data becomes our target

Read more

Summary

INTRODUCTION

The H∞ control is a robust control method which is aimed at designing the controllers to attenuate the negative effects in performance of dynamical systems caused by external disturbances guarantee the stability of systems if no disturbance exists [1]–[3]. By reviewing the existing results on H∞ control for dynamical systems, it is not difficult to find that most of researchers are concerned about model-based H∞ controller design using the variety of methods, such as linear matrix inequality (LMI) [10]–[12], zero-sum game [13]–[17] and pole assignment [18]–[20], etc. In view of the advantages of off-policy learning over on-policy learning shown in our previous result [37] wherein the off -policy Q-learning method was proposed for multi-player systems without the consideration of disturbance, developing an off-policy game Q-learning algorithm to solve H∞ control problem for discrete-time linear multi-player systems using only measured data becomes our target. The superscript T is used for the transpose. ⊗ stands for the Kronecker product. vec(L) is used to turn any matrix L into a single column vector

PROBLEM STATEMENT
QUADRATIC FORM PROOF OF VALUE
ON-POLICY GAME Q-LEARNING ALGORITHM
1: Data collection
SIMULATION RESULTS
COMPARISON RESULTS OF ON-POLICY LEARNING WITH OFF-POLICY LEARNING
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call