Abstract

AbstractThis article addresses the problem of adaptive control of nonlinear chemical processes with time‐varying dynamics. The Q‐learning and policy‐iteration algorithms have been considered in the reinforcement learning (RL) framework. The performance of the two algorithms has been tested on highly nonlinear simulated continuous stirred tank reactor (CSTR). Comparison with conventional methods shows that the RL techniques are able to achieve better performance, and robustness against uncertainties. The policy‐iteration algorithm is relatively faster in convergence, and better in robustness. Copyright © 2010 Curtin University of Technology and John Wiley & Sons, Ltd.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.