Nonparametric additive model with grouped lasso and maximizing area under the ROC curve

An ROC (Receiver Operating Characteristic) curve is a popular tool in the classification of two populations. The nonparametric additive model is used to construct a classifier which is estimated by maximizing the U-statistic type of empirical AUC (Area Under Curve). In particular, the sparsity situation is considered in the sense that only a small number of variables is significant in the classification, so it is demanded that lots of noisy variables will be removed. Some theoretical result on the necessity of variable selection under the sparsity condition is provided since the AUC of the classifier from maximization of empirical AUC is not guaranteed to be optimal. To select significant variables in the classification, the grouped lasso which has been widely used when groups of parameters need to be either selected or discarded simultaneously is used. In addition, the performance of the proposed method is evaluated by numerical studies including simulation and real data examples compared with other existing approaches.

[1]  M. Yuan,et al.  Model selection and estimation in regression with grouped variables , 2006 .

[2]  László Györfi,et al.  A Probabilistic Theory of Pattern Recognition , 1996, Stochastic Modelling and Applied Probability.

[3]  P. Billingsley,et al.  Probability and Measure , 1980 .

[4]  Tom Fawcett,et al.  Analysis and Visualization of Classifier Performance: Comparison under Imprecise Class and Cost Distributions , 1997, KDD.

[5]  Y. Li,et al.  AUC-based biomarker ensemble with an application on gene scores predicting low bone mineral density , 2011, Bioinform..

[6]  Paul Tseng,et al.  A coordinate gradient descent method for nonsmooth separable minimization , 2008, Math. Program..

[7]  T. Cai,et al.  Combining Predictors for Classification Using the Area under the Receiver Operating Characteristic Curve , 2006, Biometrics.

[8]  Kenneth Revett,et al.  Feature selection in Parkinson's disease: A rough sets approach , 2009, 2009 International Multiconference on Computer Science and Information Technology.

[9]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[10]  Todd,et al.  Diffuse large B-cell lymphoma outcome prediction by gene-expression profiling and supervised machine learning , 2002, Nature Medicine.

[11]  Jianqing Fan,et al.  Regularization of Wavelet Approximations , 2001 .

[12]  Sergey Bakin,et al.  Adaptive regression and model selection in data mining problems , 1999 .

[13]  S. Ramaswamy,et al.  Translation of microarray data into clinically relevant cancer diagnostic tests using gene expression ratios in lung cancer and mesothelioma. , 2002, Cancer research.

[14]  Jian Huang,et al.  Regularized ROC method for disease classification and biomarker selection with microarray data , 2005, Bioinform..

[15]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[16]  P. Bühlmann,et al.  The group lasso for logistic regression , 2008 .

[17]  R. Tibshirani,et al.  A note on the group lasso and a sparse group lasso , 2010, 1001.0736.

[18]  Bogdan E. Popescu,et al.  Gradient Directed Regularization for Linear Regression and Classi…cation , 2004 .

[19]  William Nick Street,et al.  Learning to Rank by Maximizing AUC with Linear Programming , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[20]  R. Tibshirani,et al.  Generalized additive models for medical research , 1986, Statistical methods in medical research.