Abstract

What are the distinct ways in which a set of predictor variables can provide information about a target variable? When does a variable provide unique information, when do variables share redundant information, and when do variables combine synergistically to provide complementary information? The redundancy lattice from the partial information decomposition of Williams and Beer provided a promising glimpse at the answer to these questions. However, this structure was constructed using a much criticised measure of redundant information, and despite sustained research, no completely satisfactory replacement measure has been proposed. In this paper, we take a different approach, applying the axiomatic derivation of the redundancy lattice to a single realisation from a set of discrete variables. To overcome the difficulty associated with signed pointwise mutual information, we apply this decomposition separately to the unsigned entropic components of pointwise mutual information which we refer to as the specificity and ambiguity. This yields a separate redundancy lattice for each component. Then based upon an operational interpretation of redundancy, we define measures of redundant specificity and ambiguity enabling us to evaluate the partial information atoms in each lattice. These atoms can be recombined to yield the sought-after multivariate information decomposition. We apply this framework to canonical examples from the literature and discuss the results and the various properties of the decomposition. In particular, the pointwise decomposition using specificity and ambiguity satisfies a chain rule over target variables, which provides new insights into the so-called two-bit-copy example.

Highlights

  • The aim of information decomposition is to divide the total amount of information provided by a set of predictor variables, about a target variable, into atoms of partial information contributed either individually or jointly by the various subsets of the predictors

  • Pointwise redundant information for a single source event ai equals the pointwise mutual information, i ∩ ai → t = i ai ; t. It seems that the step should be to define some measure of pointwise redundant information which is compatible with these pointwise partial information decomposition (PPID) axioms; there is a problem—the pointwise mutual information is not non-negative

  • The pointwise mutual information does not depend on the apportionment of the informative exclusions across the set of events which did not occur, the pointwise redundant information should not depend on this apportionment either

Read more

Summary

Introduction

The aim of information decomposition is to divide the total amount of information provided by a set of predictor variables, about a target variable, into atoms of partial information contributed either individually or jointly by the various subsets of the predictors. The problem is to define one of the unique, redundant or complementary information—something not provided by Shannon’s information theory—in order to uniquely evaluate the decomposition. The general problem of information decomposition is to provide both a structure for multivariate information which is consistent with the bivariate decomposition, and a way to uniquely evaluate the atoms in this general structure. Appendix A contains discussions regarding the so-called two-bit-copy problem in terms of Kelly gambling, Appendix B contains many of the technical details and proofs, while Appendix B contains some more examples

Notation
Partial Information Decomposition
Pointwise Information Theory
Pointwise Information Decomposition
Pointwise Unique
Pointwise Partial Information Decomposition
Two Distinct Types of Probability Mass Exclusions
The Directed Components of Pointwise Information
Operational Interpretation of Redundant Information
Motivational Example
Pointwise Partial Information Decomposition Using Specificity and Ambiguity
Bivariate PPID Using the Specificity and Ambiguity
Redundancy Measures on the Specificity and Ambiguity Lattices
Discussion
Comparison to Existing Measures
Probability Distribution X OR
Probability Distribution PWUNQ
Probability Distribution RDNERR
Probability Distribution TBC
Summary of Key Properties
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call