Rho–tau embedding and gauge freedom in information geometry

The standard model of information geometry, expressed as Fisher–Rao metric and Amari-Chensov tensor, reflects an embedding of probability density by $$\log $$log-transform. The present paper studies parametrized statistical models and the induced geometry using arbitrary embedding functions, comparing single-function approaches (Eguchi’s U-embedding and Naudts’ deformed-log or phi-embedding) and a two-function embedding approach (Zhang’s conjugate rho-tau embedding). In terms of geometry, the rho-tau embedding of a parametric statistical model defines both a Riemannian metric, called “rho-tau metric”, and an alpha-family of rho-tau connections, with the former controlled by a single function and the latter by both embedding functions $$\rho $$ρ and $$\tau $$τ in general. We identify conditions under which the rho-tau metric becomes Hessian and hence the $$\pm 1$$±1 rho-tau connections are dually flat. For any choice of rho and tau there exist models belonging to the phi-deformed exponential family for which the rho-tau metric is Hessian. In other cases the rho–tau metric may be only conformally equivalent with a Hessian metric. Finally, we show a formulation of the maximum entropy framework which yields the phi-exponential family as the solution.

[1]  Estimators, escort probabilities, and phi-exponential families in statistical physics , 2004, math-ph/0402005.

[2]  Jun Zhang,et al.  Information Geometry Under Monotone Embedding. Part I: Divergence Functions , 2017, GSI.

[3]  C. Tsallis What are the Numbers that Experiments Provide , 1994 .

[4]  Hiroshi Matsuzoe,et al.  Hessian structures on deformed exponential families and their conformal structures , 2014 .

[5]  Jun Zhang On Monotone Embedding in Information Geometry , 2015, Entropy.

[6]  Jan Naudts,et al.  The exponential family in abstract information theory , 2013, GSI.

[7]  C. Tsallis Possible generalization of Boltzmann-Gibbs statistics , 1988 .

[8]  H. Callen Thermodynamics and an Introduction to Thermostatistics , 1988 .

[9]  G. Kaniadakis,et al.  Non-linear kinetics underlying generalized statistics , 2001 .

[10]  C. Tsallis,et al.  Generalized statistical mechanics : connection with thermodynamics , 1991 .

[11]  Antonio Maria Scarfone,et al.  A Sequential Structure of Statistical Manifolds on Deformed Exponential Family , 2017, GSI.

[12]  Jan Naudts,et al.  Generalised Exponential Families and Associated Entropy Functions , 2008, Entropy.

[13]  R. Callen,et al.  Thermodynamics and an Introduction to Thermostatistics, 2nd Edition , 1985 .

[14]  Nigel J. Newton Information Geometric Nonlinear Filtering , 2015, ArXiv.

[15]  Shun-ichi Amari,et al.  Geometry of deformed exponential families: Invariant, dually-flat and conformal geometries , 2012 .

[16]  S. Eguchi Information Geometry and Statistical Pattern Recognition , 2004 .

[17]  Giovanni Pistone,et al.  Deformed Exponential Bundle: The Linear Growth Case , 2017, GSI.

[18]  C. Tsallis,et al.  The role of constraints within generalized nonextensive statistics , 1998 .

[19]  Jian Zhou Information Theory and Statistical Mechanics Revisited , 2016, ArXiv.

[20]  Shun-ichi Amari,et al.  Differential-geometrical methods in statistics , 1985 .

[21]  Jun Zhang,et al.  Divergence Function, Duality, and Convex Analysis , 2004, Neural Computation.

[22]  Jan Naudts,et al.  Information Geometry Under Monotone Embedding. Part II: Geometry , 2017, GSI.

[23]  Jun Zhang,et al.  Nonparametric Information Geometry: From Divergence Function to Referential-Representational Biduality on Statistical Manifolds , 2013, Entropy.