The problem of infinite information flow
By: Zheng Bian, Erik M. Bollt
Potential Business Impact:
Shows how much one thing affects another.
We study conditional mutual information (cMI) between a pair of variables $X,Y$ given a third one $Z$ and derived quantities including transfer entropy (TE) and causation entropy (CE) in the dynamically relevant context where $X=T(Y,Z)$ is determined by $Y,Z$ via a deterministic transformation $T$. Under mild continuity assumptions on their distributions, we prove a zero-infinity dichotomy for cMI for a wide class of $T$, which gives a yes-or-no answer to the question of information flow as quantified by TE or CE. Such an answer fails to distinguish between the relative amounts of information flow. To resolve this problem, we propose a discretization strategy and a conjectured formula to discern the \textit{relative ambiguities} of the system, which can serve as a reliable proxy for the relative amounts of information flow. We illustrate and validate this approach with numerical evidence.
Similar Papers
Testing (Conditional) Mutual Information
Data Structures and Algorithms
Finds hidden connections between three things.
Two-Parameter Rényi Information Quantities with Applications to Privacy Amplification and Soft Covering
Information Theory
Makes information math work better for privacy.
The Exploratory Study on the Relationship Between the Failure of Distance Metrics in High-Dimensional Space and Emergent Phenomena
Information Theory
Helps predict when new things will appear.