Abstract

Training deep neural networks with the error backpropagation algorithm is considered implausible from a biological perspective. Numerous recent publications suggest elaborate models for biologically plausible variants of deep learning, typically defining success as reaching around 98% test accuracy on the MNIST data set. Here, we investigate how far we can go on digit (MNIST) and object (CIFAR10) classification with biologically plausible, local learning rules in a network with one hidden layer and a single readout layer. The hidden layer weights are either fixed (random or random Gabor filters) or trained with unsupervised methods (Principal/Independent Component Analysis or Sparse Coding) that can be implemented by local learning rules. The readout layer is trained with a supervised, local learning rule. We first implement these models with rate neurons. This comparison reveals, first, that unsupervised learning does not lead to better performance than fixed random projections or Gabor filters for large hidden layers. Second, networks with localized receptive fields perform significantly better than networks with all-to-all connectivity and can reach backpropagation performance on MNIST. We then implement two of the networks – fixed, localized, random & random Gabor filters in the hidden layer – with spiking leaky integrate-and-fire neurons and spike timing dependent plasticity to train the readout layer. These spiking models achieve >98.2% test accuracy on MNIST, which is close to the performance of rate networks with one hidden layer trained with backpropagation. The performance of our shallow network models is comparable to most current biologically plausible models of deep learning. Furthermore, our results with a shallow spiking network provide an important reference and suggest the use of data sets other than MNIST for testing the performance of future models of biologically plausible deep learning.

Highlights

  • While learning a new task, synapses deep in the brain undergo task-relevant changes (Hayashi-Takagi, et al, 2015)

  • We study networks that consist of an input (l0), one hidden (l1) and an output-layer (l2) of units, connected by weight matrices W1 and W2 (Fig. 1)

  • In the biologically plausible network considered in this paper (Fig. 1b & c), the input-tohidden weights W1 are either fixed random, random Gabor filters or learned with an unsupervised method (Principal/ Independent Component Analysis or Sparse Coding)

Read more

Summary

Introduction

While learning a new task, synapses deep in the brain undergo task-relevant changes (Hayashi-Takagi, et al, 2015). These synapses are often many neurons downstream of sensors and many neurons upstream of actuators. Since the rules that govern such changes deep in the brain are poorly understood, it is appealing to draw inspiration from deep artificial neural networks (DNNs) (LeCun, Bengio, & Hinton, 2015). DNNs and the cerebral cortex share that information is processed in multiple layers of many neurons (Kriegeskorte, 2015; Yamins & DiCarlo, 2016) and that learning depends on changes of synaptic strengths (Hebb, 1949). Biological neurons communicate by sending discrete spikes as opposed to real-valued numbers used in DNNs

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call