Abstract
A new derivation is presented of maximum entropy, which is an extremizing principle for assigning probability distributions from expectation values. The additive form ΣiΦ(pi) for the maximand is first proved by requiring that, when some probabilities are given, the procedure for finding the remaining probabilities should not depend on the values of the given probabilities. This condition induces functional equations whose solution generates the additive form. To find the function Φ we assign two distributions in separate spaces from separate expectation values; then assign a joint distribution by taking these same values to be expectations of its marginals; then require these marginals to be the same as the separately assigned distributions. The resulting functional equations have only one viable solution—the entropic form Φ(z)=−z ln z. The exploitation of marginal distributions is due to Shore and Johnson [1], but the present derivation uses weaker axioms that require only consistency with the sum and product rules. In contrast to the information-theoretic derivation of Shannon [2], no interpretation of the maximand functional is involved.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.