Abstract

Audio classification has many practical applications such as noise pollution detection, wildlife monitoring, speech recognition, and more. For many of these applications, deploying classifiers on low powered devices for persistent deployment is desirable. Artificial neural networks (ANN) have achieved state-of-the-art performance on audio classification tasks; however, it is not always feasible to deploy modern ANNs to embedded devices due to their high power consumption. Biologically inspired spiking neural networks (SNN) have been shown to significantly reduce power consumption during inference when compared with equivalent ANNs, and they have also been theoretically proven to be more computationally powerful than stateless ANNs. This work proposes an audio classification system using SNNs, and a learning algorithm is developed for classification with multilayer SNNs which combines biologically plausible spike-timing-dependent plasticity (STDP) with spatial error backpropagation. By allowing the STDP process to account for both temporal dependencies and the non-differentiable activation function derivative, the proposed learning rule successfully trains multilayer SNNs for the considered classification tasks. Through the STDP process, the proposed learning rule is also capable of online learning; explicit storage of values from previous timesteps is not required, in contrast to the widely adopted backpropagation through time (BPTT) algorithm. The proposed approach approaches the performance of SNNs trained via BPTT on the classification tasks. SNNs trained with the proposed learning rule are evaluated on the Iris Flower dataset, the Real-World Computing Partnership sounds dataset, the Free Spoken Digits Dataset, and the UrbanSound8k dataset.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call