Abstract

Information theory has been extensively applied to neuroscience problems. The mutual information between input and output has been postulated as an objective, which neuronal systems may optimize. However, only recently the energy efficiency has been addressed within an information-theoretic framework [1]. Here, the key idea is to consider capacity per unit cost (measured in bits per joule, bpj) as the objective. We are interested in how biologically plausible constraints affect predictions made by this new theory for bpj-maximizing model neurons. More specifically, in our contribution, in line with [1] and [2], a neuron is modeled as a memory-less constant communication channel with a Gamma conditional probability distribution function (PDF) [1]. In this setting, the channel input and output are the excitatory postsynaptic potential intensity, λ, and the inter spike interval (ISI), t, with PDFs f (λ) and fT (t), respectively. We then formulate two new constraints: First, we impose a lower bound tmin on the duration tof ISIs. The rational for this is to account for a maximal firing rate. Second, we consider a peak energy expenditure constraint per ISI as compared to only bounding the expected energy expenditure. This translates into an upper bound tmax on the ISI duration. We then derive the fT (t) (corresponding to valid f (λ)) of a bpj-maximizing neuron for the original unconstrained setting from [1] and in the presence of the above two constraints for different expected ISIs. (Details omitted here for brevity.) Figure 1 shows three fT (t)s obtained in the unconstrained (dashed curves) and constrained settings (solid curves) for tmin = 1and tmax = 5. While the constrained and unconstrained solutions have the same mean, the shape of their fT (t) differ. For comparison

Highlights

  • Information theory has been extensively applied to neuroscience problems

  • In our contribution, in line with [1] and [2], a neuron is modeled as a memory-less constant communication channel with a Gamma conditional probability distribution function (PDF) [1]

  • The CV is predicted i) to be lower in the constrained setting, and ii) to increase and decrease with the mean inter spike interval (ISI) while it only decreases in the unconstrained setting

Read more

Summary

Introduction

Information theory has been extensively applied to neuroscience problems. The mutual information between input and output has been postulated as an objective, which neuronal systems may optimize. In our contribution, in line with [1] and [2], a neuron is modeled as a memory-less constant communication channel with a Gamma conditional probability distribution function (PDF) [1]. The channel input and output are the excitatory postsynaptic potential intensity, λ, and the inter spike interval (ISI), t, with PDFs f (λ) and fT (t), respectively.

Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call