Generalized Symmetric Divergence Measures and Inequalities

There are three classical divergence measures known in the literature on information theory and statistics. These are namely, Jeffryes-Kullback-Leiber \cite{jef} \cite{kul} \textit{J-divergence}. Sibson-Burbea-Rao \cite{sib} \cite{bur1, bur2} \textit{Jensen-Shannon divegernce}and Taneja \cite{tan3} \textit{Arithemtic-Geometric divergence}. These three measures bears an interesting relationship among each other. The divergence measures like \textit{Hellinger discrimination}, \textit{symmetric}$\chi ^2 - $\textit{divergence}, and \textit{triangular discrimination} are also known in the literature. All these measures can be written as particular cases of Csisz\'{a}r's \textit{f-divergence}. Recently, author proved an inequality relating all the six measures. In this paper our aim is to give one parametric generalizations of the above measures and established relationships among them. A new measure similar to \textit{Hellinger's} and \textit{triangular discriminations} is also derived.

[1]  I. J. Taneja Relative Divergence Measures and Information Inequalities , 2005, math/0505204.

[2]  I. J. Taneja Bounds on Non - Symmetric Divergence Measures in terms of Symmetric Divergence Measures - communica , 2005, math/0506256.

[3]  I. J. Taneja ON A DIFFERENCE OF JENSEN INEQUALITY AND ITS APPLICATIONS TO MEAN DIVERGENCE MEASURES , 2005, math/0501302.

[4]  Inder Jeet Taneja,et al.  Relative information of type s, Csiszár's f-divergence, and information inequalities , 2004, Inf. Sci..

[5]  I. J. Taneja Generalized Relative Information and Information Inequalities , 2003 .

[6]  New Approximations for f-Divergence via Trapezoid and Midpoint Inequalities , 2002 .

[7]  Huaiyu Zhu On Information and Sufficiency , 1997 .

[8]  I. J. Taneja New Developments in Generalized Information Measures , 1995 .

[9]  Jianhua Lin,et al.  Divergence measures based on the Shannon entropy , 1991, IEEE Trans. Inf. Theory.

[10]  Inder Jeet Taneja,et al.  On Generalized Information Measures and Their Applications , 1989 .

[11]  L. L. Cam,et al.  Asymptotic Methods In Statistical Decision Theory , 1986 .

[12]  Timothy R. C. Read,et al.  Multinomial goodness-of-fit tests , 1984 .

[13]  C. R. Rao,et al.  Entropy differential metric, distance and divergence measures in probability spaces: A unified approach , 1982 .

[14]  C. R. Rao,et al.  On the convexity of some divergence measures based on entropy functions , 1982, IEEE Trans. Inf. Theory.

[15]  K. Stolarsky,et al.  Generalizations of the Logarithmic Mean , 1975 .

[16]  I. Vajda On thef-divergence and singularity of probability measures , 1972 .

[17]  R. Sibson Information radius , 1969 .

[18]  H. Jeffreys An invariant form for the prior probability in estimation problems , 1946, Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences.

[19]  E. Hellinger,et al.  Neue Begründung der Theorie quadratischer Formen von unendlichvielen Veränderlichen. , 1909 .

[20]  K. Pearson On the Criterion that a Given System of Deviations from the Probable in the Case of a Correlated System of Variables is Such that it Can be Reasonably Supposed to have Arisen from Random Sampling , 1900 .