From the Publisher: One hundred years ago, the fundamental building block of the central nervous system, the neuron, was discovered. This study focuses the existing mathematical models of neurons and their interactions, the simulation of which has been one of the biggest challenges facing modern science. More than fifty years ago, W. S. McCulloch and W. Pitts devised their model for the neuron, John von Neumann seemed to sense the possibilities for the development of intelligent systems, and Frank Rosenblatt came up with a functioning network of neurons. Despite these advances, the subject had begun to fade as a major research area until John Hopfield arrived the scene. Drawing an analogy between neural networks and the Ising spin models of ferromagnetism, Hopfield was able to introduce a computational that would decline toward stable minima under the operation of the system of neurodynamics devised by Roy Glauber. Like a switch, a neuron is said to be either on or off. The state of the neuron is determined by the states of the other neurons and the connections between them, and the connections are assumed to be reciprocal - that is, neuron number one influences neuron number two exactly as strongly as neuron number two influences neuron number one. According to the Glauber dynamics, the states of the neurons are updated in a random serial way until an equilibrium is reached. An energy function can be associated with each state, and equilibrium corresponds to a minimum of this energy. It follows from Hopfield's assumption of reciprocity that an equilibrium will always be reached. D. H. Ackley, G. E. Hinton, and T. J. Sejnowski modified the Hopfield network by introducing the simulated annealing algorithm to search out the deepest minima. This is accomplished by - loosely speaking - shaking the machine. The violence of the shaking is controlled by a parameter called temperature, producing the Boltzmann machine - a name designed to emphasize the connection to the
Read full abstract