Abstract
This paper presents a method to improve the conservative (shape-independent) controllers from guaranteed-cost LMIs using Q-learning policy iteration setups which learn optimal controllers from data. In this context, the proposed approach uses an initial parametrization based on the LMI solution and the learning process includes parameters related to the nonlinearity of the system (and its gradient). An approximation of the Q-function using polynomials of the membership functions in Takagi-Sugeno models is obtained using a policy iteration algorithm. The resulting controller is shape-dependent, that is, uses the knowledge of membership functions, their gradients and data to improve well-known LMI solutions.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.