Hypothesis Testing, Information Divergence and Computational Geometry

In this paper, we consider the Bayesian multiple hypothesis testing problem from the stance of computational geometry. We first recall that the probability of error of the optimal decision rule, the maximum a posteriori probability (MAP) criterion, is related to both the total variation and the Chernoff statistical distances. We then consider the exponential family manifolds, and show that the MAP rule amounts to a nearest neighbor classifier that can be implemented either by point locations in an additive Bregman Voronoi diagram or by nearest neighbor queries using various techniques of computational geometry. Finally, we show that computing the best error exponent upper bounding the probability of error, the Chernoff distance, amounts to (1) find a unique geodesic/bisector intersection point for binary hypothesis, (2) solve a closest Bregman pair problem for multiple hypothesis.

[1]  Martin E. Hellman,et al.  Probability of error, equivocation, and the Chernoff bound , 1970, IEEE Trans. Inf. Theory.

[2]  Frank Nielsen,et al.  Tailored Bregman Ball Trees for Effective Nearest Neighbors , 2009 .

[3]  Don H. Johnson,et al.  On the asymptotics of M-hypothesis Bayesian detection , 1997, IEEE Trans. Inf. Theory.

[4]  M. Brandon Westover,et al.  Asymptotic Geometry of Multiple Hypothesis Testing , 2008, IEEE Transactions on Information Theory.

[5]  J. Boissonnat,et al.  Algorithmic Geometry: Frontmatter , 1998 .

[6]  Frank Nielsen,et al.  An Information-Geometric Characterization of Chernoff Information , 2013, IEEE Signal Processing Letters.

[7]  Frank Nielsen,et al.  K-MLE: A fast algorithm for learning statistical mixture models , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[8]  Richard Nock,et al.  On Bregman Voronoi diagrams , 2007, SODA '07.

[9]  David R. Bull,et al.  Robust texture features for blurred images using Undecimated Dual-Tree Complex Wavelets , 2014, 2014 IEEE International Conference on Image Processing (ICIP).

[10]  Frank Nielsen,et al.  The Burbea-Rao and Bhattacharyya Centroids , 2010, IEEE Transactions on Information Theory.

[11]  Frank Nielsen,et al.  K-nearest neighbor search: Fast GPU-based implementations and application to high-dimensional feature matching , 2010, 2010 IEEE International Conference on Image Processing.

[12]  L. Brown Fundamentals of statistical exponential families: with applications in statistical decision theory , 1986 .

[13]  Frank Nielsen,et al.  Bregman vantage point trees for efficient nearest Neighbor Queries , 2009, 2009 IEEE International Conference on Multimedia and Expo.

[14]  Inderjit S. Dhillon,et al.  Clustering with Bregman Divergences , 2005, J. Mach. Learn. Res..

[15]  Shun-ichi Amari,et al.  Methods of information geometry , 2000 .

[16]  H. Chernoff A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations , 1952 .

[17]  Mariette Yvinec,et al.  Algorithmic geometry , 1998 .