Abstract

It is critical for biological neural networks to recover their functions when some neurons of networks are damaged by lesion or aging. Self-organized criticality which induced neuronal avalanche whose distribution of size following a power-law [1] provides computational optimality to neural networks [2]. We show that the criticality of neural networks, an intrinsic property of complex networks, also provides the robustness of computational capability to neural networks. We constructed a neural network which consisted of 300 integrate-and-fire neurons with random connections. For neural networks to exhibit self-organized criticality, we used a dynamic synapse model, neural network with activity-dependent depressive synapses, which followed the Liquid State Machine Paradigm [3]. We used EEG recordings obtained during a two-class classification task from The Wadsworth BCI Dataset (IIb) [4]. EEG signals of only 4 channels were chosen for inputs to neural networks. Thus, two readout units were connected to neural networks which determined one of the classes from input signals. We estimated the classification performance of these data for 4 states of the neural network: initial random state (A state), critical state (B state), damaged state (C state, i.e. 10% neurons removed) and critical state of the damaged neural network (D state). To confirm that neural network can transit into critical states, we followed the procedure based on the paper of Levina et al. [1]. We found that the features of self-organized criticality or power-law distribution in the damaged neural networks could be maintained by the dynamic synapses as shown in Figure ​Figure1.1. The classification performance was recovered from the damaged state when neural networks were into critical states (Table ​(Table1).1). The root mean square error (RMSE) increased in the C state, but the criticality of neural networks reduced errors to the level of a undamaged state. These results demonstrate that the criticality may contribute to recovery for capability of computation in damaged neural networks. Figure 1 The distribution of the avalanches size in the neural network. Table 1 Root Mean Square Errors (RMSE) of each state in the neural network

Highlights

  • It is critical for biological neural networks to recover their functions when some neurons of networks are damaged by lesion or aging

  • We constructed a neural network which consisted of 300 integrate-and-fire neurons with random connections

  • We used EEG recordings obtained during a two-class classification task from The Wadsworth BCI Dataset (IIb) [4]

Read more

Summary

Introduction

It is critical for biological neural networks to recover their functions when some neurons of networks are damaged by lesion or aging. We constructed a neural network which consisted of 300 integrate-and-fire neurons with random connections.

Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.