Abstract

In the training of artificial neural networks, reinforcement learning substitutes a qualitative binary target of “success” or “failure” for the quantitative error criterion of supervised learning. By this method of learning control can be established for the special case of operation in which no objective function exists. If no model whatsoever exists of the dynamics of a chemical process, it still may be possible to train an artificial neural network to control the process much as a human being would learn by trial and error. We describe a network architecture for process control, and explain quantitatively how the weights on the connections in the network can be adjusted to yield the desired control action. An example of a nonlinear CSTR is used to illustrate the performance of the proposed net and compare it with that of a tuned PID controller. The net can be trained to meet almost any criteria selected for control as long as the criteria can be expressed in the form of inequality constraints, but requires extensive, and perhaps excessively long, training on a serial computer to do so.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call