Abstract
One measure of the computational power of a single unit in a network is the rate at which it can transfer information. Because that rate depends both on the unit's analog resolution and on its speed, many types of units may reach their optimum information flow when they operate very fast, with only coarse or binary resolution at each time step (this occurs because the information rate scales linearly with speed or bandwidth, but only logarithmically with analog resolution). Furthermore, calculations for a mean-square-limited channel with Gaussian noise show that the optimum information rate for the best possible analog code is only about 30% better than the optimum rate for the simplest binary code. For the biologically realistic case of an irregular neural spike train, a binary code carries about two orders of magnitude more information than an analog rate-code using the same spikes, and is equally consistent with neurophysiology. Thus, neural networks that optimally transfer information would probably use a binary (or coarse analog) code, rather than the high-resolution, real-valued codes usually explored.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.