Partial Information Decomposition
   HOME

TheInfoList



OR:

Partial Information Decomposition is an extension of
information theory Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
, that aims to generalize the pairwise relations described by information theory to the interaction of multiple variables.


Motivation

Information theory can quantify the amount of information a single source variable X_1 has about a target variable Y via the
mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
I(X_1;Y). If we now consider a second source variable X_2, classical information theory can only describe the mutual information of the joint variable \ with Y, given by I(X_1,X_2;Y). In general however, it would be interesting to know how exactly the individual variables X_1 and X_2 and their interactions relate to Y. Consider that we are given two source variables X_1, X_2 \in \ and a target variable Y=XOR(X_1,X_2). In this case the total mutual information I(X_1,X_2;Y)=1, while the individual mutual information I(X_1;Y)=I(X_2;Y)=0. That is, there is
synergistic Synergy is an interaction or cooperation giving rise to a whole that is greater than the simple sum of its parts (i.e., a non-linear addition of force, energy, or effect). The term ''synergy'' comes from the Attic Greek word συνεργία ' f ...
information arising from the interaction of X_1,X_2 about Y, which cannot be easily captured with classical information theoretic quantities.


Definition

Partial information decomposition further decomposes the mutual information between the source variables \ with the target variable Y as I(X_1,X_2;Y)=\text(X_1;Y \setminus X_2) + \text(X_2;Y \setminus X_1) + \text(X_1,X_2;Y) + \text(X_1,X_2;Y) Here the individual information atoms are defined as * \text(X_1;Y \setminus X_2) is the ''unique'' information that X_1 has about Y, which is not in X_2 * \text(X_1,X_2;Y) is the ''synergistic'' information that is in the interaction of X_1 and X_2 about Y * \text(X_1,X_2;Y) is the ''redundant'' information that is in both X_1 or X_2 about Y There is, thus far, no universal agreement on how these terms should be defined, with different approaches that decompose information into redundant, unique, and synergistic components appearing in the literature.


Applications

Despite the lack of universal agreement, partial information decomposition has been applied to diverse fields, including climatology, neuroscience sociology, and machine learning Partial information decomposition has also been proposed as a possible foundation on which to build a mathematically robust definition of emergence in complex systems and may be relevant to formal theories of consciousness.


See also

*
Mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
*
Total correlation In probability theory and in particular in information theory, total correlation (Watanabe 1960) is one of several generalizations of the mutual information. It is also known as the ''multivariate constraint'' (Garner 1962) or ''multiinformation'' ...
*
Dual total correlation In information theory, dual total correlation, information rate, excess entropy,Nihat Ay, E. Olbrich, N. Bertschinger (2001). A unifying framework for complexity measures of finite systems. European Conference on Complex Systemspdf or binding inf ...
* Interaction information


References

{{Reflist Information theory