Abstract

Information theory provides robust measures of multivariable interdependence, but classically does little to characterize the multivariable relationships it detects. The Partial Information Decomposition (PID) characterizes the mutual information between variables by decomposing it into unique, redundant, and synergistic components. This has been usefully applied, particularly in neuroscience, but there is currently no generally accepted method for its computation. Independently, the Information Delta framework characterizes non-pairwise dependencies in genetic datasets. This framework has developed an intuitive geometric interpretation for how discrete functions encode information, but lacks some important generalizations. This paper shows that the PID and Delta frameworks are largely equivalent. We equate their key expressions, allowing for results in one framework to apply towards open questions in the other. For example, we find that the approach of Bertschinger et al. is useful for the open Information Delta question of how to deal with linkage disequilibrium. We also show how PID solutions can be mapped onto the space of delta measures. Using Bertschinger et al. as an example solution, we identify a specific plane in delta-space on which this approach’s optimization is constrained, and compute it for all possible three-variable discrete functions of a three-letter alphabet. This yields a clear geometric picture of how a given solution decomposes information.

Highlights

  • The variables in complex biological data frequently have nonlinear and non-pairwise dependency relationships

  • An analytical approach formulated by Williams and Beer frames these questions in terms of the Partial Information Decomposition (PID) [1]

  • We suggest an approach for the analysis of genetic datasets which would return both the closest discrete function underlying the data and its PID in the Bertschinger solution, and which would require no further optimization after the initial construction of a solution library

Read more

Summary

Introduction

The variables in complex biological data frequently have nonlinear and non-pairwise dependency relationships. We show that the sets of probability distributions, Q, used by Bertschinger can be mapped onto low-dimensional manifolds in ~δ-space, which intersect with the ~δ-plane of Figure 2 This approach is theoretically useful for the Delta information framework, since it factors out X, Y dependence in the data, thereby accounting for linkage disequilibrium between genetic variables. We suggest an approach for the analysis of genetic datasets which would return both the closest discrete function underlying the data and its PID in the Bertschinger solution, and which would require no further optimization after the initial construction of a solution library This realization yields a low-dimensional geometric interpretation of this optimization problem, and we compute the solution for all possible three-variable discrete functions of alphabet size three. Code to replicate these computations and the associated figures is freely available [20]

Interaction Information and Multi-Information
Information Decomposition
Information Deltas and Their Geometry
Information Decomposition in Terms of Deltas
Relationship between Diagonal and Interaction Information
The Function Plane
Transforming Probability Tensors within Q
PID Calculation for All Functions
Alternate Solutions
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call