Abstract

Abstract In this paper, we show that an information source composed with n random variables may be split into 2 n or 2 n − 1 “states”; therefore, one could compute the maximum entropy of the source. We derive the efficiency and the unused capacity of an information source. We demonstrate that in more than two dimensions, the transmission's variability depends on the system configuration; thus, we determine the upper and the lower bounds to the mutual information and propose the transmission power as an indicator of the Triple Helix of university–industry–government relationships. The transmission power is defined as the fraction of the total ‘configurational information’ produced in a system; it appears like the efficiency of the transmission and may be interpreted as the strength of the variables dependency, the strength of the synergy between the system's variable or the strength of information flow within the system.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.