Generalized informational functionals and new monotone measures of statistical complexity
By: Razvan Gabriel Iagar, David Puertas-Centeno
Potential Business Impact:
Finds new ways to measure information in data.
In this paper we introduce a biparametric family of transformations which can be seen as an extension of the so-called up and down transformations. This new class of transformations allows to us to introduce new informational functionals, which we have called \textit{down-moments} and \textit{cumulative upper-moments}. A remarkable fact is that the down-moments provide, in some cases, an interpolation between the $p$-th moments and the power R\'enyi entropies of a probability density. We establish new and sharp inequalities relating these new functionals to the classical informational measures such as moments, R\'enyi and Shannon entropies and Fisher information measures. We also give the optimal bounds as well as the minimizing densities, which are in some cases expressed in terms of the generalized trigonometric functions. We furthermore define new classes of measures of statistical complexity obtained as quotients of the new functionals, and establish monotonicity properties for them through an algebraic conjugation of up and down transformations. All of these properties highlight an intricate structure of functional inequalities.
Similar Papers
Two-Parameter Rényi Information Quantities with Applications to Privacy Amplification and Soft Covering
Information Theory
Makes information math work better for privacy.
Generalized Derangetropy Functionals for Modeling Cyclical Information Flow
Information Theory
Helps computers understand how information changes over time.
Simple and Sharp Generalization Bounds via Lifting
Statistics Theory
Makes computer learning more accurate and faster.