The analytic dually flat space of the mixture family of two prescribed distinct Cauchy distributions

A smooth and strictly convex function on an open convex domain induces both (1) a Hessian manifold with respect to the standard flat Euclidean connection, and (2) a dually flat space of information geometry. We first review these constructions and illustrate how to instantiate them for (a) full regular exponential families from their partition functions, (b) regular homogeneous cones from their characteristic functions, and (c) mixture families from their Shannon negentropy functions. Although these structures can be explicitly built for many common examples of the first two classes, the differential entropy of a continuous statistical mixture with distinct prescribed density components sharing the same support is hitherto not known in closed form, hence forcing implementations of mixture family manifolds in practice using Monte Carlo sampling. In this work, we report a notable exception: The family of mixtures defined as the convex combination of two prescribed and distinct Cauchy distributions. As a byproduct, we report closed-form formula for the Jensen-Shannon divergence between two mixtures of two prescribed Cauchy components.

[1]  Zhenning Zhang,et al.  Information geometry of the power inverse Gaussian distribution , 2007 .

[2]  Hirohiko Shima,et al.  Geometry of Hessian Structures , 2013, GSI.

[3]  Frank Nielsen,et al.  Geometry and Fixed-Rate Quantization in Riemannian Metric Spaces Induced by Separable Bregman Divergences , 2019, GSI.

[4]  Jean-Pierre Crouzeix,et al.  A relationship between the second derivatives of a convex function and of its conjugate , 1977, Math. Program..

[5]  Inderjit S. Dhillon,et al.  Clustering with Bregman Divergences , 2005, J. Mach. Learn. Res..

[6]  L. Bregman The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming , 1967 .

[7]  André F. T. Martins,et al.  Learning with Fenchel-Young Losses , 2020, J. Mach. Learn. Res..

[8]  Frank Nielsen,et al.  Monte Carlo Information-Geometric Structures , 2018, Geometric Structures of Information.

[9]  Huafei Sun,et al.  THE GEOMETRY OF THE DIRICHLET MANIFOLD , 2008 .

[10]  O. Barndorff-Nielsen Information and Exponential Families in Statistical Theory , 1980 .

[11]  Frank Nielsen,et al.  A closed-form formula for the Kullback-Leibler divergence between Cauchy distributions , 2019, ArXiv.

[12]  Osman Güler,et al.  Barrier Functions in Interior Point Methods , 1996, Math. Oper. Res..

[13]  Giovanni Pistone,et al.  Information Geometry of the Gaussian Distribution in View of Stochastic Optimization , 2015, FOGA.

[14]  S. Eguchi Geometry of minimum contrast , 1992 .

[15]  M. Frei Analysis On Symmetric Cones , 2016 .

[16]  J. Naudts,et al.  DATA SET MODELS AND EXPONENTIAL FAMILIES IN STATISTICAL PHYSICS AND BEYOND , 2011, 1111.0490.

[17]  F. Opitz Information geometry and its applications , 2012, 2012 9th European Radar Conference.

[18]  Frank Nielsen,et al.  An Elementary Introduction to Information Geometry , 2018, Entropy.

[19]  R. Rockafellar Conjugates and Legendre Transforms of Convex Functions , 1967, Canadian Journal of Mathematics.

[20]  Frank Nielsen,et al.  The Burbea-Rao and Bhattacharyya Centroids , 2010, IEEE Transactions on Information Theory.

[21]  Frank Nielsen,et al.  On $f$-divergences between Cauchy distributions , 2021, GSI.

[22]  Man-Chung Yue,et al.  Universal Barrier Is n-Self-Concordant , 2018, Math. Oper. Res..

[23]  Christopher T. J. Dodson,et al.  Introduction to Riemannian Geometry , 2008 .

[24]  Keisuke Yamazaki,et al.  Kullback Information of Normal Mixture is not an Analytic Function , 2004 .

[25]  Jianhua Lin,et al.  Divergence measures based on the Shannon entropy , 1991, IEEE Trans. Inf. Theory.

[26]  Frank Nielsen,et al.  On the Geometry of Mixtures of Prescribed Distributions , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[27]  Julianna Pinele,et al.  The Fisher–Rao Distance between Multivariate Normal Distributions: Special Cases, Bounds and Applications , 2020, Entropy.

[28]  C. Atkinson Rao's distance measure , 1981 .

[29]  Richard Nock,et al.  On Bregman Voronoi diagrams , 2007, SODA '07.

[30]  F. Nielsen On geodesic triangles with right angles in a dually flat space , 2019, Signals and Communication Technology.

[31]  S. Amari,et al.  Information geometry of divergence functions , 2010 .