On Optimal Pairwise Linear Classifiers for Normal Distributions: The Two-Dimensional Case

Optimal Bayesian linear classifiers have been studied in the literature for many decades. We demonstrate that all the known results consider only the scenario when the quadratic polynomial has coincident roots. Indeed, we present a complete analysis of the case when the optimal classifier between two normally distributed classes is pairwise and linear. We focus on some special cases of the normal distribution with nonequal covariance matrices. We determine the conditions that the mean vectors and covariance matrices have to satisfy in order to obtain the optimal pairwise linear classifier. As opposed to the state of the art, in all the cases discussed here, the linear classifier is given by a pair of straight lines, which is a particular case of the general equation of second degree. We also provide some empirical results, using synthetic data for the Minsky's paradox case, and demonstrated that the linear classifier achieves very good performance. Finally, we have tested our approach on real life data obtained from the UCI machine learning repository. The empirical results that we obtained show the superiority of our scheme over the traditional Fisher's discriminant classifier.

[1]  E. Deeba,et al.  Interactive Linear Algebra with Maple V , 1998 .

[2]  Thien M. Ha Optimum Decision Rules in Pattern Recognition , 1998, SSPR/SPR.

[3]  Sarunas Raudys,et al.  Evolution and generalization of a single neurone. III. Primitive, regularized, standard, robust and minimax regressions , 2000, Neural Networks.

[4]  Sarunas Raudys,et al.  Evolution and generalization of a single neurone: I. Single-layer perceptron as seven statistical classifiers , 1998, Neural Networks.

[5]  Andrew R. Webb,et al.  Statistical Pattern Recognition , 1999 .

[6]  Keinosuke Fukunaga,et al.  Introduction to statistical pattern recognition (2nd ed.) , 1990 .

[7]  W. V. McCarthy,et al.  Discriminant Analysis with Singular Covariance Matrices: Methods and Applications to Spectroscopic Data , 1995 .

[8]  Sarunas Raudys,et al.  Evolution and generalization of a single neurone: : II. Complexity of statistical classifiers and sample size considerations , 1998, Neural Networks.

[9]  O. J. Murphy,et al.  Nearest neighbor pattern classification perceptrons , 1990, Proc. IEEE.

[10]  Sarunas Raudys,et al.  On Dimensionality, Sample Size, and Classification Error of Nonparametric Linear Classification Algorithms , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Thien M. Ha,et al.  The Optimum Class-Selective Rejection Rule , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Mayer Aladjem Linear Discriminant Analysis for Two Classes via Removal of Classification Structure , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Peter E. Hart,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[14]  B. John Oommen,et al.  On optimal pairwise linear classifiers for normal distributions: the d-dimensional case , 2003, Pattern Recognit..

[15]  David G. Stork,et al.  Pattern Classification , 1973 .

[16]  William Nick Street,et al.  Breast Cancer Diagnosis and Prognosis Via Linear Programming , 1995, Oper. Res..

[17]  Keinosuke Fukunaga,et al.  Introduction to Statistical Pattern Recognition , 1972 .

[18]  Yoshua Bengio,et al.  Pattern Recognition and Neural Networks , 1995 .

[19]  Kenneth Rose,et al.  A Deterministic Annealing Approach for Parsimonious Design of Piecewise Regression Models , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[20]  Josef Kittler,et al.  Pattern recognition : a statistical approach , 1982 .

[21]  Vwani P. Roychowdhury,et al.  An Adaptive Stochastic Approximation Algorithm for Simultaneous Diagonalization of Matrix Sequences With Applications , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[22]  J. T. Brown,et al.  The elements of analytical geometry , 1934 .

[23]  Witold Malina,et al.  On an Extended Fisher Criterion for Feature Selection , 1981, IEEE Transactions on Pattern Analysis and Machine Intelligence.