Neyman-Pearson classification under a strict constraint

Motivated by problems of anomaly detection, this paper implements the Neyman-Pearson paradigm to deal with asymmetric errors in binary classication with a convex loss. Given a nite collection of classiers, we combine them and obtain a new classier that satises simultaneously the two following properties with high probability: (i), its probability of type I error is below a pre-specied level and (ii), it has probability of type II error close to the minimum possible. The proposed classier is obtained by minimizing an empirical objective subject to an empirical constraint. The novelty of the method is that the classier output by this problem is shown to satisfy the original constraint on type I error. This strict enforcement of the constraint has interesting consequences on the control of the type II error and we develop new techniques to handle this situation. Finally, connections with chance constrained optimization are evident and are investigated.

[1]  S. Boucheron,et al.  Theory of classification : a survey of some recent advances , 2005 .

[2]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[3]  David Casasent,et al.  Radial basis function neural networks for nonlinear Fisher discrimination and Neyman-Pearson classification , 2003, Neural Networks.

[4]  E. Lehmann Testing Statistical Hypotheses , 1960 .

[5]  V. Koltchinskii,et al.  Oracle inequalities in empirical risk minimization and sparse recovery problems , 2011 .

[6]  Robert D. Nowak,et al.  A Neyman-Pearson approach to statistical learning , 2005, IEEE Transactions on Information Theory.

[7]  Gilles Blanchard,et al.  Semi-Supervised Novelty Detection , 2010, J. Mach. Learn. Res..

[8]  Xiang Li,et al.  Probabilistically Constrained Linear Programs and Risk-Adjusted Controller Design , 2005, SIAM J. Optim..

[9]  Giuseppe Carlo Calafiore,et al.  The scenario approach to robust control design , 2006, IEEE Transactions on Automatic Control.

[10]  Xin Tong,et al.  Neyman-Pearson Classification, Convexity and Stochastic Constraints , 2011, J. Mach. Learn. Res..

[11]  K. N. Dollman,et al.  - 1 , 1743 .

[12]  R. Schapire The Strength of Weak Learnability , 1990, Machine Learning.

[13]  Clayton Scott,et al.  Performance Measures for Neyman–Pearson Classification , 2007, IEEE Transactions on Information Theory.

[14]  Michael I. Jordan,et al.  Convexity, Classification, and Risk Bounds , 2006 .

[15]  László Györfi,et al.  A Probabilistic Theory of Pattern Recognition , 1996, Stochastic Modelling and Applied Probability.

[16]  Alexander Shapiro,et al.  Convex Approximations of Chance Constrained Programs , 2006, SIAM J. Optim..

[17]  Zhaoxu Sun,et al.  Analysis to Neyman-Pearson classification with convex loss function , 2008 .