Event Abstract Back to Event Layered sparse associative network for soft pattern classification and contextual pattern completion Evan Ehrenberg1*, Pentti Kanerva1 and Friedrich Sommer1 1 University of California at Berkeley, Redwood Center for Theoretical Neuroscience, United States Traditional models of associative networks [1, 5] have impressive pattern completion capabilities. They feature high storage capacity and can reconstruct incomplete sensory input from stored memory in a context-sensitive fashion. However, associative networks are not useful as pattern classifiers beyond simple toy examples where the classes form well-separated clusters in pattern space. Conversely, multi-layer feed-forward neural networks [4] can be trained to solve challenging classification tasks, but they cannot perform pattern completion. Here we tested the ability of sparse two-layer associative networks to learn to 1) recognize patterns in real world data and 2) use memory content to reconstruct noisy input in a context-sensitive fashion. Specifically, we used the memory models with supervised learning on handwritten characters (NIST database [3]) and assessed how well the memory could predict the class of unknown input and reconstruct the input pattern. We started with the Kanerva memory model [2], which has a two-layer structure. The first neural layer has fixed random synapses and contains many more neurons than input fibers, which are sparsely activated. This stage maps the input space into a high-dimensional space of sparse patterns. The synapses of second layer neurons store associations between the high-dimensional sparse patterns and desired output patterns via Hebbian plasticity. We trained the model with the NIST data by storing for each input pattern the given digit interpretation as well as the input pattern in autoassociative fashion. When cross-validated with (noisy) inputs, the model frequently performed pattern recognition and reconstruction successfully. However, ambiguous inputs could not be classified, and the reconstruction was a mixture pattern not corresponding to a valid handwritten digit. To overcome these limitations, we constructed a new model with a two-layer structure similar to the described Kanerva model but with two important new features, enabling soft classification and pattern completion in two subsequent phases. One new feature is labeling first layer neurons during supervised training according to how their activity is correlated with the occurrence of certain classes. Thus, in the classification phase when a new input drives first layer neurons, estimates of class membership are encoded in active populations of first layer cells. The second feature is a phase of selective pattern completion performed after classification. In this phase first layer neurons are also influenced by lateral interactions and feedback from the second layer. This competitive dynamical process accomplishes pattern completion that is contingent on a specific interpretation rather than creating a mixture of the entire memory content, even if the input is ambiguous. We compare the model to current methods of classifying the NIST data. We found that the model exhibits competitive classification performance and adds a valuable explanatory component not provided by ordinary classification algorithms. Finally, we discuss how the proposed memory architecture would map onto dentate gyrus and CA3 of the hippocampus.
Read full abstract