In
probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
information theory
Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
, the interaction information is a generalization of the
mutual information
In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
for more than two
variables
Variable may refer to:
Computer science
* Variable (computer science), a symbolic name associated with a value and whose associated value may be changed
Mathematics
* Variable (mathematics), a symbol that represents a quantity in a mathemat ...
.
There are many names for interaction information, including ''amount of information'', ''information correlation'', ''co-information'',
and simply ''mutual information''.
Interaction information expresses the amount of information (redundancy or synergy) bound up in a set of variables, ''beyond'' that which is present in any subset of those variables. Unlike the mutual information, the interaction information can be either positive or negative. These functions, their negativity and minima have a direct interpretation in
algebraic topology
Algebraic topology is a branch of mathematics that uses tools from abstract algebra to study topological spaces. The basic goal is to find algebraic invariant (mathematics), invariants that classification theorem, classify topological spaces up t ...
.
Definition
The
conditional mutual information
Conditional (if then) may refer to:
*Causal conditional, if X then Y, where X is a cause of Y
*Conditional probability, the probability of an event A given that another event B
*Conditional proof, in logic: a proof that asserts a conditional, a ...
can be used to inductively define the interaction information for any finite number of variables as follows:
:
where
:
Some authors define the interaction information differently, by swapping the two terms being subtracted in the preceding equation. This has the effect of reversing the sign for an odd number of variables.
For three variables
, the interaction information
is given by
:
where
is the mutual information between variables
and
, and
is the
conditional mutual information
Conditional (if then) may refer to:
*Causal conditional, if X then Y, where X is a cause of Y
*Conditional probability, the probability of an event A given that another event B
*Conditional proof, in logic: a proof that asserts a conditional, a ...
between variables
and
given
. The interaction information is
symmetric
Symmetry () in everyday life refers to a sense of harmonious and beautiful proportion and balance. In mathematics, the term has a more precise definition and is usually used to refer to an object that is invariant under some transformations ...
, so it does not matter which variable is conditioned on. This is easy to see when the interaction information is written in terms of entropy and joint entropy, as follows:
:
In general, for the set of variables
, the interaction information can be written in the following form (compare with
Kirkwood approximation):
:
For three variables, the interaction information measures the influence of a variable
on the amount of information shared between
and
. Because the term
can be larger than
, the interaction information can be negative as well as positive. This will happen, for example, when
and
are independent but not conditionally independent given
. Positive interaction information indicates that variable
inhibits (i.e., ''accounts for'' or ''explains'' some of) the correlation between
and
, whereas negative interaction information indicates that variable
facilitates or enhances the correlation.
Properties
Interaction information is bounded. In the three variable case, it is bounded by
:
If three variables form a Markov chain
, then
, but
. Therefore
:
Examples
Positive interaction information
Positive interaction information seems much more natural than negative interaction information in the sense that such ''explanatory'' effects are typical of common-cause structures. For example, clouds cause rain and also block the sun; therefore, the correlation between rain and darkness is partly accounted for by the presence of clouds,
. The result is positive interaction information
.
Negative interaction information
A car's engine can fail to start due to either a dead battery or a blocked fuel pump. Ordinarily, we assume that battery death and fuel pump blockage are independent events,
. But knowing that the car fails to start, if an inspection shows the battery to be in good health, we can conclude that the fuel pump must be blocked. Therefore
, and the result is negative interaction information.
Difficulty of interpretation
The possible negativity of interaction information can be the source of some confusion.
Many authors have taken zero interaction information as a sign that three or more random variables do not interact, but this interpretation is wrong.
To see how difficult interpretation can be, consider a set of eight independent
binary variables . Agglomerate these variables as follows:
:
Because the
's overlap each other (are redundant) on the three binary variables
, we would expect the interaction information
to equal
bits, which it does. However, consider now the agglomerated variables
:
These are the same variables as before with the addition of
. However,
in this case is actually equal to
bit, indicating less redundancy. This is correct in the sense that
:
but it remains difficult to interpret.
Uses
* Jakulin and Bratko (2003b) provide a machine learning algorithm which uses interaction information.
* Killian, Kravitz and Gilson (2007) use mutual information expansion to extract entropy estimates from molecular simulations.
* LeVine and Weinstein (2014) use interaction information and other N-body information measures to quantify allosteric couplings in molecular simulations.
* Moore et al. (2006), Chanda P, Zhang A, Brazeau D, Sucheston L, Freudenheim JL, Ambrosone C, Ramanathan M. (2007) and Chanda P, Sucheston L, Zhang A, Brazeau D, Freudenheim JL, Ambrosone C, Ramanathan M. (2008) demonstrate the use of interaction information for analyzing gene-gene and gene-environmental interactions associated with complex diseases.
* Pandey and Sarkar (2017) use interaction information in Cosmology to study the influence of large-scale environments on galaxy properties.
* A python package for computing all multivariate interaction or mutual informations, conditional mutual information, joint entropies, total correlations, information distance in a dataset of n variables is available .
See also
*
Mutual information
In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
*
Total correlation In probability theory and in particular in information theory, total correlation (Watanabe 1960) is one of several generalizations of the mutual information. It is also known as the ''multivariate constraint'' (Garner 1962) or ''multiinformation'' ...
*
Dual total correlation
In information theory, dual total correlation, information rate, excess entropy,Nihat Ay, E. Olbrich, N. Bertschinger (2001). A unifying framework for complexity measures of finite systems. European Conference on Complex Systemspdf or binding inf ...
*
Partial Information Decomposition Partial Information Decomposition is an extension of information theory, that aims to generalize the pairwise relations described by information theory to the interaction of multiple variables.
Motivation
Information theory can quantify the amo ...
References
*
* Bell, A J (2003), ''The co-information lattice'
* Fano, R M (1961), ''Transmission of Information: A Statistical Theory of Communications'', MIT Press, Cambridge, MA.
* Garner W R (1962). ''Uncertainty and Structure as Psychological Concepts'', JohnWiley & Sons, New York.
*
*
* Hu Kuo Tin (1962), On the Amount of Information. Theory Probab. Appl.,7(4), 439-44
PDF* Jakulin A & Bratko I (2003a). Analyzing Attribute Dependencies, in N Lavra\quad, D Gamberger, L Todorovski & H Blockeel, eds, ''Proceedings of the 7th European Conference on Principles and Practice of Knowledge Discovery in Databases'', Springer, Cavtat-Dubrovnik, Croatia, pp. 229–240.
* Jakulin A & Bratko I (2003b). Quantifying and visualizing attribute interaction
*
*
* Moore JH, Gilbert JC, Tsai CT, Chiang FT, Holden T, Barney N, White BC (2006). A flexible computational framework for detecting, characterizing, and interpreting statistical patterns of epistasis in genetic studies of human disease susceptibility, ''Journal of Theoretical Biology'' 241, 252-261
* Nemenman I (2004). Information theory, multivariate dependence, and genetic network inferenc
* Pearl, J (1988), ''Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference'', Morgan Kaufmann, San Mateo, CA.
* Tsujishita, T (1995), On triple mutual information, ''Advances in applied mathematics'' 16, 269-274.
*
*
*
* LeVine MV, Weinstein H (2014), NbIT - A New Information Theory-Based Analysis of Allosteric Mechanisms Reveals Residues that Underlie Function in the Leucine Transporter LeuT. ''PLoS Computational Biology''
* {{cite journal , last1 = Pandey , first1 = Biswajit , last2 = Sarkar , first2 = Suman , year = 2017 , title = How much a galaxy knows about its large-scale environment?: An information theoretic perspective , journal = Monthly Notices of the Royal Astronomical Society Letters, volume = 467 , issue = 1, page = L6 , doi=10.1093/mnrasl/slw250, doi-access = free , arxiv = 1611.00283, bibcode = 2017MNRAS.467L...6P, s2cid = 119095496
* https://www3.nd.edu/~jnl/ee80653/Fall2005/tutorials/sunil.pdf
* Yeung R W (1992). A new outlook on Shannon's information measures. in IEEE Transactions on Information Theory
Information theory